[ 488.358437] env[61907]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61907) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.358879] env[61907]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61907) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.358879] env[61907]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61907) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 488.359203] env[61907]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 488.455225] env[61907]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61907) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 488.465316] env[61907]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61907) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 489.081034] env[61907]: INFO nova.virt.driver [None req-ab8b16e7-84b8-43e3-b715-f34b5272f9eb None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 489.153800] env[61907]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 489.154289] env[61907]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 489.154556] env[61907]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61907) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 492.386962] env[61907]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-f46b6ddd-07d1-41e8-9fcb-8a90f2c92233 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.403642] env[61907]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61907) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 492.403774] env[61907]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-4f9830ad-748d-401c-b59d-441a2ba82190 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.436122] env[61907]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 9d0cd. [ 492.436257] env[61907]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.282s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 492.436796] env[61907]: INFO nova.virt.vmwareapi.driver [None req-ab8b16e7-84b8-43e3-b715-f34b5272f9eb None None] VMware vCenter version: 7.0.3 [ 492.440229] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67438aec-a23a-4145-b4ab-d496578cd752 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.457090] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53e07be5-a500-4ad0-ab52-b6dc19f2be0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.463388] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4478fde9-9436-404d-a4d7-0c5b5f09e807 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.469828] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83163ef1-bed6-4c03-bcd0-8a7686ea53f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.482489] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15db4f3-2fc9-4fb6-ba29-a0a03018bc87 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.488325] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692aaf79-26da-458b-ac84-3ccc1d5bb2c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.518243] env[61907]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-1465eaa2-f971-4fdc-babe-4e988b19ab3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 492.522766] env[61907]: DEBUG nova.virt.vmwareapi.driver [None req-ab8b16e7-84b8-43e3-b715-f34b5272f9eb None None] Extension org.openstack.compute already exists. {{(pid=61907) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:227}} [ 492.525394] env[61907]: INFO nova.compute.provider_config [None req-ab8b16e7-84b8-43e3-b715-f34b5272f9eb None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 493.028568] env[61907]: DEBUG nova.context [None req-ab8b16e7-84b8-43e3-b715-f34b5272f9eb None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),2483c22e-cb9e-4609-804b-ea6089a3b4bb(cell1) {{(pid=61907) load_cells /opt/stack/nova/nova/context.py:464}} [ 493.030791] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.031057] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.031735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.032186] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Acquiring lock "2483c22e-cb9e-4609-804b-ea6089a3b4bb" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.032378] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Lock "2483c22e-cb9e-4609-804b-ea6089a3b4bb" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.033391] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Lock "2483c22e-cb9e-4609-804b-ea6089a3b4bb" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 493.053488] env[61907]: INFO dbcounter [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Registered counter for database nova_cell0 [ 493.061907] env[61907]: INFO dbcounter [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Registered counter for database nova_cell1 [ 493.065159] env[61907]: DEBUG oslo_db.sqlalchemy.engines [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61907) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.065519] env[61907]: DEBUG oslo_db.sqlalchemy.engines [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61907) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 493.070356] env[61907]: ERROR nova.db.main.api [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.070356] env[61907]: result = function(*args, **kwargs) [ 493.070356] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 493.070356] env[61907]: return func(*args, **kwargs) [ 493.070356] env[61907]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.070356] env[61907]: result = fn(*args, **kwargs) [ 493.070356] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.070356] env[61907]: return f(*args, **kwargs) [ 493.070356] env[61907]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 493.070356] env[61907]: return db.service_get_minimum_version(context, binaries) [ 493.070356] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.070356] env[61907]: _check_db_access() [ 493.070356] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.070356] env[61907]: stacktrace = ''.join(traceback.format_stack()) [ 493.070356] env[61907]: [ 493.071421] env[61907]: ERROR nova.db.main.api [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 493.071421] env[61907]: result = function(*args, **kwargs) [ 493.071421] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 493.071421] env[61907]: return func(*args, **kwargs) [ 493.071421] env[61907]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 493.071421] env[61907]: result = fn(*args, **kwargs) [ 493.071421] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 493.071421] env[61907]: return f(*args, **kwargs) [ 493.071421] env[61907]: File "/opt/stack/nova/nova/objects/service.py", line 556, in _db_service_get_minimum_version [ 493.071421] env[61907]: return db.service_get_minimum_version(context, binaries) [ 493.071421] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 493.071421] env[61907]: _check_db_access() [ 493.071421] env[61907]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 493.071421] env[61907]: stacktrace = ''.join(traceback.format_stack()) [ 493.071421] env[61907]: [ 493.071836] env[61907]: WARNING nova.objects.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 493.071949] env[61907]: WARNING nova.objects.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Failed to get minimum service version for cell 2483c22e-cb9e-4609-804b-ea6089a3b4bb [ 493.072394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Acquiring lock "singleton_lock" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 493.072556] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Acquired lock "singleton_lock" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 493.072796] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Releasing lock "singleton_lock" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 493.073137] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Full set of CONF: {{(pid=61907) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 493.073282] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ******************************************************************************** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 493.073409] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Configuration options gathered from: {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 493.073569] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 493.073740] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 493.073864] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ================================================================================ {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 493.074085] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] allow_resize_to_same_host = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.074261] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] arq_binding_timeout = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.074395] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] backdoor_port = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.074520] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] backdoor_socket = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.074685] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] block_device_allocate_retries = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.074850] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] block_device_allocate_retries_interval = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075028] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cert = self.pem {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075199] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075368] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute_monitors = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075534] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] config_dir = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075705] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] config_drive_format = iso9660 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.075838] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076008] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] config_source = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076183] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] console_host = devstack {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076347] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] control_exchange = nova {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076503] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cpu_allocation_ratio = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076661] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] daemon = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076826] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] debug = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.076981] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_access_ip_network_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.077166] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_availability_zone = nova {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.077322] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_ephemeral_format = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.077482] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_green_pool_size = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.077744] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.077921] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] default_schedule_zone = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078096] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] disk_allocation_ratio = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078262] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] enable_new_services = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078441] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] enabled_apis = ['osapi_compute'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078607] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] enabled_ssl_apis = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078803] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] flat_injected = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.078969] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] force_config_drive = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.079148] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] force_raw_images = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.079321] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] graceful_shutdown_timeout = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.079485] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] heal_instance_info_cache_interval = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.079699] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] host = cpu-1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.079880] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080054] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] initial_disk_allocation_ratio = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080220] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] initial_ram_allocation_ratio = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080472] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080592] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_build_timeout = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080751] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_delete_interval = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.080954] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_format = [instance: %(uuid)s] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.081164] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_name_template = instance-%08x {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.081333] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_usage_audit = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.081506] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_usage_audit_period = month {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.081675] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.081845] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] instances_path = /opt/stack/data/nova/instances {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082020] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] internal_service_availability_zone = internal {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082180] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] key = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082342] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] live_migration_retry_count = 30 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082509] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_color = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082676] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_config_append = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.082846] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083012] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_dir = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083186] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083316] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_options = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083478] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_rotate_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083646] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_rotate_interval_type = days {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083814] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] log_rotation_type = none {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.083943] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084078] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084251] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084418] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084547] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084712] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] long_rpc_timeout = 1800 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.084893] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_concurrent_builds = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085094] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_concurrent_live_migrations = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085262] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_concurrent_snapshots = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085423] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_local_block_devices = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085583] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_logfile_count = 30 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085741] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] max_logfile_size_mb = 200 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.085923] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] maximum_instance_delete_attempts = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.086140] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metadata_listen = 0.0.0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.086323] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metadata_listen_port = 8775 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.086493] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metadata_workers = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.086653] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] migrate_max_retries = -1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.086822] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] mkisofs_cmd = genisoimage {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087037] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] my_block_storage_ip = 10.180.1.21 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087173] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] my_ip = 10.180.1.21 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087378] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087543] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] network_allocate_retries = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087750] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.087927] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] osapi_compute_listen = 0.0.0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088113] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] osapi_compute_listen_port = 8774 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088288] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] osapi_compute_unique_server_name_scope = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088458] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] osapi_compute_workers = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088623] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] password_length = 12 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088838] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] periodic_enable = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.088942] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] periodic_fuzzy_delay = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089125] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] pointer_model = usbtablet {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089296] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] preallocate_images = none {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089457] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] publish_errors = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089591] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] pybasedir = /opt/stack/nova {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089753] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ram_allocation_ratio = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.089917] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rate_limit_burst = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090101] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rate_limit_except_level = CRITICAL {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090265] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rate_limit_interval = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090441] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reboot_timeout = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090604] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reclaim_instance_interval = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090764] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] record = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.090934] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reimage_timeout_per_gb = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091116] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] report_interval = 120 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091279] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rescue_timeout = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091438] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reserved_host_cpus = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091598] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reserved_host_disk_mb = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091760] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reserved_host_memory_mb = 512 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.091922] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] reserved_huge_pages = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092099] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] resize_confirm_window = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092262] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] resize_fs_using_block_device = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092422] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] resume_guests_state_on_host_boot = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092589] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092750] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] rpc_response_timeout = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.092913] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] run_external_periodic_tasks = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.093158] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] running_deleted_instance_action = reap {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.093337] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] running_deleted_instance_poll_interval = 1800 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.093498] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] running_deleted_instance_timeout = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.093660] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler_instance_sync_interval = 120 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.093830] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_down_time = 720 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094000] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] servicegroup_driver = db {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094174] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] shell_completion = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094336] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] shelved_offload_time = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094493] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] shelved_poll_interval = 3600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094664] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] shutdown_timeout = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094829] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] source_is_ipv6 = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.094991] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ssl_only = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.095251] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.095422] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] sync_power_state_interval = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.095610] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] sync_power_state_pool_size = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.095801] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] syslog_log_facility = LOG_USER {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.095965] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] tempdir = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096146] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] timeout_nbd = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096316] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] transport_url = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096479] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] update_resources_interval = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096642] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_cow_images = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096807] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_eventlog = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.096966] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_journal = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097144] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_json = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097308] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_rootwrap_daemon = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097468] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_stderr = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097628] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] use_syslog = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097815] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vcpu_pin_set = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.097992] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plugging_is_fatal = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.098179] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plugging_timeout = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.098351] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] virt_mkfs = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.098513] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] volume_usage_poll_interval = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.098677] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] watch_log_file = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.098847] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] web = /usr/share/spice-html5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 493.099047] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.099219] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.099387] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.099560] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_concurrency.disable_process_locking = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.099846] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100049] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100223] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100396] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100569] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100736] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.100922] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.auth_strategy = keystone {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101105] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.compute_link_prefix = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101287] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101467] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.dhcp_domain = novalocal {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101639] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.enable_instance_password = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101808] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.glance_link_prefix = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.101976] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.102168] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.102348] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.instance_list_per_project_cells = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.102514] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.list_records_by_skipping_down_cells = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.102680] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.local_metadata_per_cell = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.102852] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.max_limit = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103029] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.metadata_cache_expiration = 15 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103208] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.neutron_default_tenant_id = default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103382] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.response_validation = warn {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103552] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.use_neutron_default_nets = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103722] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.103889] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104070] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104249] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104421] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_dynamic_targets = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104586] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_jsonfile_path = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104771] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.104969] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.backend = dogpile.cache.memcached {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.105153] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.backend_argument = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.105328] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.config_prefix = cache.oslo {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.105499] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.dead_timeout = 60.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.105684] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.debug_cache_backend = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.105867] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.enable_retry_client = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106047] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.enable_socket_keepalive = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106224] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.enabled = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106394] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.enforce_fips_mode = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106560] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.expiration_time = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106726] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.hashclient_retry_attempts = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.106897] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.hashclient_retry_delay = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107075] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_dead_retry = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107241] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_password = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107409] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107572] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107762] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_pool_maxsize = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.107944] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108126] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_sasl_enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108310] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108481] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_socket_timeout = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108646] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.memcache_username = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108817] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.proxies = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.108985] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_db = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.109164] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_password = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.109335] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_sentinel_service_name = mymaster {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.109513] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.109684] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_server = localhost:6379 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.109856] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_socket_timeout = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110033] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.redis_username = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110203] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.retry_attempts = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110371] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.retry_delay = 0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110535] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.socket_keepalive_count = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110703] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.socket_keepalive_idle = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.110868] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.socket_keepalive_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111037] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.tls_allowed_ciphers = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111202] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.tls_cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111363] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.tls_certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111527] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.tls_enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111687] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cache.tls_keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.111867] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112053] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.auth_type = password {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112223] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112401] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.catalog_info = volumev3::publicURL {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112562] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112727] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.112893] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.cross_az_attach = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113067] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.debug = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113234] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.endpoint_template = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113401] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.http_retries = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113565] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113727] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.113898] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.os_region_name = RegionOne {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114079] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114242] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cinder.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114418] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114579] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.cpu_dedicated_set = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114739] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.cpu_shared_set = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.114906] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.image_type_exclude_list = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115079] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115246] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.max_concurrent_disk_ops = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115408] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.max_disk_devices_to_attach = -1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115571] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115762] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.115937] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.resource_provider_association_refresh = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116112] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116279] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.shutdown_retry_interval = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116461] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116642] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] conductor.workers = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116826] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] console.allowed_origins = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.116988] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] console.ssl_ciphers = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.117175] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] console.ssl_minimum_version = default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.117348] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] consoleauth.enforce_session_timeout = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.117516] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] consoleauth.token_ttl = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.117717] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.117891] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118072] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118265] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118398] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118558] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118744] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.118911] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119092] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119257] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119418] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119577] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119738] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.119939] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.service_type = accelerator {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120133] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120299] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120477] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120644] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120830] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.120994] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] cyborg.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.121192] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.backend = sqlalchemy {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.121366] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.connection = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.121535] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.connection_debug = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.121711] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.connection_parameters = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.121876] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.connection_recycle_time = 3600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122056] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.connection_trace = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122227] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.db_inc_retry_interval = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122393] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.db_max_retries = 20 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122558] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.db_max_retry_interval = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122722] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.db_retry_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.122888] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.max_overflow = 50 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123062] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.max_pool_size = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123231] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.max_retries = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123402] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123583] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.mysql_wsrep_sync_wait = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123743] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.pool_timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.123909] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.retry_interval = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124082] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.slave_connection = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124253] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.sqlite_synchronous = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124414] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] database.use_db_reconnect = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124593] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.backend = sqlalchemy {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124762] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.connection = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.124928] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.connection_debug = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125115] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.connection_parameters = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125280] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.connection_recycle_time = 3600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125443] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.connection_trace = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125605] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.db_inc_retry_interval = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125783] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.db_max_retries = 20 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.125950] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.db_max_retry_interval = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126127] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.db_retry_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126292] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.max_overflow = 50 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126455] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.max_pool_size = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126619] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.max_retries = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126793] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.126953] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.127125] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.pool_timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.127289] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.retry_interval = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.127448] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.slave_connection = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.127614] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] api_database.sqlite_synchronous = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.127827] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] devices.enabled_mdev_types = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128026] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128207] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ephemeral_storage_encryption.default_format = luks {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128374] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ephemeral_storage_encryption.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128540] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128740] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.api_servers = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.128918] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129096] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129267] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129430] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129587] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129749] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.debug = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.129918] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.default_trusted_certificate_ids = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130092] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.enable_certificate_validation = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130260] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.enable_rbd_download = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130419] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130584] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130745] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.130906] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131075] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131241] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.num_retries = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131411] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.rbd_ceph_conf = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131574] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.rbd_connect_timeout = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131741] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.rbd_pool = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.131911] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.rbd_user = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132081] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132244] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132402] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132570] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.service_type = image {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132738] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.132898] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133070] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133233] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133432] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133602] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.verify_glance_signatures = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133764] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] glance.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.133932] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] guestfs.debug = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134119] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134286] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134465] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134627] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134793] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.134957] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135136] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135296] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135462] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135623] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135785] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.135949] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136124] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136285] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136445] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136619] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.service_type = shared-file-system {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136787] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.share_apply_policy_timeout = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.136952] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137127] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137289] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137449] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137640] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137827] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] manila.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.137999] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] mks.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.138378] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.138574] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.manager_interval = 2400 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.138770] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.precache_concurrency = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.138950] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.remove_unused_base_images = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.139138] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.139312] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.139490] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] image_cache.subdirectory_name = _base {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.139670] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.api_max_retries = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.139841] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.api_retry_interval = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140014] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140190] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140354] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140515] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140684] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.140850] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.conductor_group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141026] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141210] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141396] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141565] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141745] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.141929] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142111] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142286] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.peer_list = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142462] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142629] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142818] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.serial_console_state_timeout = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.142983] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143172] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.service_type = baremetal {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143336] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.shard = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143502] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143663] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143825] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.143985] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.144180] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.144346] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ironic.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.144530] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.144705] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] key_manager.fixed_key = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.144889] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145065] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.barbican_api_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145228] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.barbican_endpoint = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145402] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.barbican_endpoint_type = public {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145565] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.barbican_region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145728] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.145889] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146064] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146231] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146392] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146555] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.number_of_retries = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146719] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.retry_delay = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.146882] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.send_service_user_token = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147057] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147222] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147386] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.verify_ssl = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147548] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican.verify_ssl_path = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147742] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.147922] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148097] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148261] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148425] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148586] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148758] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.148922] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149094] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] barbican_service_user.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149266] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.approle_role_id = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149428] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.approle_secret_id = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149600] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.kv_mountpoint = secret {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149763] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.kv_path = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.149931] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.kv_version = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150131] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.namespace = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150302] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.root_token_id = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150467] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.ssl_ca_crt_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150640] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.timeout = 60.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150808] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.use_ssl = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.150978] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151172] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151339] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151502] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151663] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151829] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.151988] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152162] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152322] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152485] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152645] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152804] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.152962] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153133] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153294] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153455] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153629] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.service_type = identity {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153798] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.153960] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.154136] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.154298] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.154478] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.154642] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] keystone.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.154844] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.connection_uri = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155017] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_mode = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155189] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_model_extra_flags = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155361] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_models = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155531] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_power_governor_high = performance {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155701] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_power_governor_low = powersave {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.155864] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_power_management = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156042] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156212] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.device_detach_attempts = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156374] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.device_detach_timeout = 20 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156539] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.disk_cachemodes = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156699] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.disk_prefix = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.156873] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.enabled_perf_events = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157050] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.file_backed_memory = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157224] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.gid_maps = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157383] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.hw_disk_discard = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157541] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.hw_machine_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157747] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_rbd_ceph_conf = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.157932] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158114] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158288] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_rbd_glance_store_name = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158456] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_rbd_pool = rbd {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158627] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_type = default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158789] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.images_volume_group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.158952] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.inject_key = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159131] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.inject_partition = -2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159295] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.inject_password = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159458] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.iscsi_iface = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159621] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.iser_use_multipath = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159793] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_bandwidth = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.159952] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.160175] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_downtime = 500 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.160348] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.160515] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.160678] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_inbound_addr = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.160874] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161064] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_permit_post_copy = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161231] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_scheme = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161407] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_timeout_action = abort {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161569] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_tunnelled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161730] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_uri = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.161894] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.live_migration_with_native_tls = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.162065] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.max_queues = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.162233] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.162470] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.162636] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.nfs_mount_options = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.162930] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163118] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163287] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_iser_scan_tries = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163448] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_memory_encrypted_guests = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163612] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163795] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_pcie_ports = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.163981] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.num_volume_scan_tries = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.164167] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.pmem_namespaces = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.164331] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.quobyte_client_cfg = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.164620] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.164797] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rbd_connect_timeout = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.164966] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165145] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165308] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rbd_secret_uuid = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165468] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rbd_user = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165635] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165806] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.remote_filesystem_transport = ssh {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.165970] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rescue_image_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.166143] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rescue_kernel_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.166304] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rescue_ramdisk_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.166475] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.166637] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.rx_queue_size = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.166854] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.smbfs_mount_options = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.167179] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.167361] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.snapshot_compression = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.167527] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.snapshot_image_format = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.167771] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.167949] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.sparse_logical_volumes = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168129] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.swtpm_enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168301] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.swtpm_group = tss {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168471] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.swtpm_user = tss {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168641] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.sysinfo_serial = unique {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168803] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.tb_cache_size = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.168961] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.tx_queue_size = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.169143] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.uid_maps = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.169310] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.use_virtio_for_bridges = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.169480] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.virt_type = kvm {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.169651] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.volume_clear = zero {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.169839] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.volume_clear_size = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170054] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.volume_use_multipath = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170232] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_cache_path = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170406] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170576] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_mount_group = qemu {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170745] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_mount_opts = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.170914] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171212] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171394] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.vzstorage_mount_user = stack {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171563] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171738] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.171913] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.auth_type = password {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172087] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172252] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172416] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172576] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172735] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.172935] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.default_floating_pool = public {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173121] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173290] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.extension_sync_interval = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173454] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.http_retries = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173617] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173786] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.173939] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174122] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174286] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174455] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.ovs_bridge = br-int {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174621] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.physnets = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174789] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.region_name = RegionOne {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.174954] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175136] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.service_metadata_proxy = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175298] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175468] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.service_type = network {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175633] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175814] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.175978] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176151] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176396] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176584] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] neutron.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176763] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] notifications.bdms_in_notifications = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.176947] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] notifications.default_level = INFO {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177141] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] notifications.notification_format = unversioned {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177312] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] notifications.notify_on_state_change = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177493] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177689] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] pci.alias = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.177879] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] pci.device_spec = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178062] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] pci.report_in_placement = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178242] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178415] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.auth_type = password {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178583] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178756] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.178934] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179114] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179278] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179438] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179598] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.default_domain_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179758] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.default_domain_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.179918] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.domain_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180115] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.domain_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180285] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180448] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180608] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180765] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.180925] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181113] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.password = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181277] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.project_domain_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181446] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.project_domain_name = Default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181987] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.project_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181987] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.project_name = service {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.181987] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.region_name = RegionOne {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182131] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182266] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182438] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.service_type = placement {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182603] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182763] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.182924] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183094] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.system_scope = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183256] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183417] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.trust_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183575] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.user_domain_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183744] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.user_domain_name = Default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.183906] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.user_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184090] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.username = nova {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184277] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184440] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] placement.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184618] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.cores = 20 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184782] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.count_usage_from_placement = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.184953] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185150] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.injected_file_content_bytes = 10240 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185322] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.injected_file_path_length = 255 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185492] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.injected_files = 5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185663] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.instances = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185831] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.key_pairs = 100 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.185996] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.metadata_items = 128 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186192] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.ram = 51200 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186359] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.recheck_quota = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186528] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.server_group_members = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186695] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] quota.server_groups = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.186867] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187042] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187209] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.image_metadata_prefilter = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187373] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187539] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.max_attempts = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187784] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.max_placement_results = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.187902] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188103] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.query_placement_for_image_type_support = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188277] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188452] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] scheduler.workers = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188631] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.188817] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189025] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189203] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189372] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189537] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189702] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.189892] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190116] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.host_subset_size = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190286] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190448] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190614] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190784] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.isolated_hosts = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.190953] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.isolated_images = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191132] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191296] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191459] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191621] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.pci_in_placement = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191796] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.191964] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192145] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192321] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192485] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192648] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.192841] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.track_instance_changes = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193040] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193220] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metrics.required = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193450] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metrics.weight_multiplier = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193640] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.193819] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] metrics.weight_setting = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194155] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194336] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194514] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.port_range = 10000:20000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194685] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.194856] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195035] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] serial_console.serialproxy_port = 6083 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195208] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195381] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.auth_type = password {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195542] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195700] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.195866] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196038] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196200] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196370] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.send_service_user_token = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196536] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196695] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] service_user.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.196869] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.agent_enabled = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197041] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197356] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197550] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197751] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.html5proxy_port = 6082 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.197944] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.image_compression = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198124] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.jpeg_compression = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198288] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.playback_compression = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198450] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.require_secure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198619] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.server_listen = 127.0.0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198806] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.198978] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.streaming_mode = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199152] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] spice.zlib_compression = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199320] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] upgrade_levels.baseapi = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199490] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] upgrade_levels.compute = auto {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199652] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] upgrade_levels.conductor = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199810] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] upgrade_levels.scheduler = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.199990] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200187] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200351] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200511] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200673] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.200861] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201033] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201201] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201359] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vendordata_dynamic_auth.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201534] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.api_retry_count = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201696] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.ca_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.201894] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.cache_prefix = devstack-image-cache {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202078] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.cluster_name = testcl1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202250] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.connection_pool_size = 10 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202408] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.console_delay_seconds = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202577] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.datastore_regex = ^datastore.* {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202781] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.202954] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.host_password = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203136] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.host_port = 443 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203309] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.host_username = administrator@vsphere.local {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203474] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.insecure = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203635] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.integration_bridge = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203800] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.maximum_objects = 100 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.203959] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.pbm_default_policy = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204134] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.pbm_enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204295] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.pbm_wsdl_location = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204464] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204625] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.serial_port_proxy_uri = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204783] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.serial_port_service_uri = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.204951] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.task_poll_interval = 0.5 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205137] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.use_linked_clone = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205312] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.vnc_keymap = en-us {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205478] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.vnc_port = 5900 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205648] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vmware.vnc_port_total = 10000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.205838] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.auth_schemes = ['none'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206027] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206327] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206516] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206694] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.novncproxy_port = 6080 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.206874] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.server_listen = 127.0.0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207060] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207225] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.vencrypt_ca_certs = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207386] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.vencrypt_client_cert = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207548] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vnc.vencrypt_client_key = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207759] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.207945] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_deep_image_inspection = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208137] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208305] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208468] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208694] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.disable_rootwrap = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.208879] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.enable_numa_live_migration = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209057] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209226] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209390] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209556] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.libvirt_disable_apic = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209716] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.209880] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210076] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210252] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210431] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210576] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210738] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.210899] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211077] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211247] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211432] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211602] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.client_socket_timeout = 900 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211767] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.default_pool_size = 1000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.211930] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.keep_alive = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212105] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.max_header_line = 16384 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212306] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.secure_proxy_ssl_header = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212423] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.ssl_ca_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212579] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.ssl_cert_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212737] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.ssl_key_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.212899] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.tcp_keepidle = 600 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213089] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213281] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] zvm.ca_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.213446] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] zvm.cloud_connector_url = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214178] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214385] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] zvm.reachable_timeout = 300 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214578] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.enforce_new_defaults = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.214957] env[61907]: WARNING oslo_config.cfg [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 493.215162] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.enforce_scope = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215345] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.policy_default_rule = default {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215534] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215719] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.policy_file = policy.yaml {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.215891] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216065] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216231] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216390] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216555] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216726] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.216906] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217096] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.connection_string = messaging:// {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217269] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.enabled = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217440] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.es_doc_type = notification {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217604] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.es_scroll_size = 10000 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217800] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.es_scroll_time = 2m {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.217972] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.filter_error_trace = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218156] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.hmac_keys = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218327] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.sentinel_service_name = mymaster {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218494] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.socket_timeout = 0.1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218658] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.trace_requests = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.218824] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler.trace_sqlalchemy = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219013] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler_jaeger.process_tags = {} {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219183] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler_jaeger.service_name_prefix = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219348] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] profiler_otlp.service_name_prefix = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219513] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] remote_debug.host = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219675] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] remote_debug.port = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.219856] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220047] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220245] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220428] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220599] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220773] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.220937] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221112] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221303] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221481] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221645] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221817] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.221986] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222174] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222346] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222518] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222679] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.222852] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223026] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223193] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223357] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223530] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223691] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.223858] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224027] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224192] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224353] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224516] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224683] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.224850] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225035] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225214] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225376] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225546] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225715] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.ssl_version = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.225879] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226078] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226249] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_notifications.retry = -1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226433] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226609] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_messaging_notifications.transport_url = **** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226783] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.auth_section = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.226946] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.auth_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227118] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.cafile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227280] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.certfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227443] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.collect_timing = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227603] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.connect_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227792] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.connect_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.227956] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.endpoint_id = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228132] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.endpoint_override = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228295] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.insecure = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228450] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.keyfile = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228606] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.max_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228781] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.min_version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.228955] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.region_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229134] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.retriable_status_codes = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229296] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.service_name = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229455] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.service_type = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229616] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.split_loggers = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229773] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.status_code_retries = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.229934] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.status_code_retry_delay = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230135] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.timeout = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230302] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.valid_interfaces = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230461] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_limit.version = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230648] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_reports.file_event_handler = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230788] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.230951] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] oslo_reports.log_dir = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231139] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231301] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231460] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231627] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231793] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.231956] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232192] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232380] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232566] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232713] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.232879] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233053] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] vif_plug_ovs_privileged.user = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233231] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.flat_interface = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233411] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233588] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233761] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.233934] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234117] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234289] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234454] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234633] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234804] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.isolate_vif = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.234975] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235155] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235330] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235502] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.ovsdb_interface = native {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235670] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] os_vif_ovs.per_port_bridge = False {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.235844] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.capabilities = [21] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236014] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236181] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.helper_command = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236348] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236512] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236674] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] privsep_osbrick.user = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.236850] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237034] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.group = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237234] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.helper_command = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237422] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237588] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237773] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] nova_sys_admin.user = None {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 493.237912] env[61907]: DEBUG oslo_service.service [None req-89b91ffc-2fe5-49ba-a895-2262e1a6a51a None None] ******************************************************************************** {{(pid=61907) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 493.238427] env[61907]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 493.741557] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Getting list of instances from cluster (obj){ [ 493.741557] env[61907]: value = "domain-c8" [ 493.741557] env[61907]: _type = "ClusterComputeResource" [ 493.741557] env[61907]: } {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 493.742847] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1362e64d-df4f-408f-9e68-a88552a1552f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 493.751658] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Got total of 0 instances {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 493.752205] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 493.752672] env[61907]: INFO nova.virt.node [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Generated node identity 91bca385-a423-4ca4-9da0-aeb4615e22d3 [ 493.752908] env[61907]: INFO nova.virt.node [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Wrote node identity 91bca385-a423-4ca4-9da0-aeb4615e22d3 to /opt/stack/data/n-cpu-1/compute_id [ 494.256322] env[61907]: WARNING nova.compute.manager [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Compute nodes ['91bca385-a423-4ca4-9da0-aeb4615e22d3'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 495.262119] env[61907]: INFO nova.compute.manager [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 496.268265] env[61907]: WARNING nova.compute.manager [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 496.268614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.268821] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.268969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 496.269149] env[61907]: DEBUG nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 496.270087] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae0b2c8-6916-4326-b945-081f91c613ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.278466] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529619cc-fdd0-4f13-b051-fcebd5118a49 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.293322] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ffa70f-3f85-4d38-bae7-4c1b9a22dddc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.299596] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d053c64f-c5a9-49d0-bf67-598e5f5229d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.327317] env[61907]: DEBUG nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181526MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 496.327460] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.327650] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 496.830459] env[61907]: WARNING nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] No compute node record for cpu-1:91bca385-a423-4ca4-9da0-aeb4615e22d3: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 91bca385-a423-4ca4-9da0-aeb4615e22d3 could not be found. [ 497.334334] env[61907]: INFO nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 91bca385-a423-4ca4-9da0-aeb4615e22d3 [ 498.845716] env[61907]: DEBUG nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 498.845716] env[61907]: DEBUG nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 499.001422] env[61907]: INFO nova.scheduler.client.report [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] [req-3e29dee6-6bf0-4fd2-b501-8cae47a83893] Created resource provider record via placement API for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 499.015566] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc9768c-dc1a-457f-ae39-130a2a043ad3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.024136] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02fb6a23-0f68-4b11-9514-886374800d07 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.053337] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b095c4-4ec0-421b-9895-f6e847b2c63f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.061032] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef9cb5c-d318-4335-ba60-8906fc214e9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.074046] env[61907]: DEBUG nova.compute.provider_tree [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.614020] env[61907]: DEBUG nova.scheduler.client.report [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 499.614020] env[61907]: DEBUG nova.compute.provider_tree [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 0 to 1 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 499.614020] env[61907]: DEBUG nova.compute.provider_tree [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 499.664104] env[61907]: DEBUG nova.compute.provider_tree [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 1 to 2 during operation: update_traits {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 500.169396] env[61907]: DEBUG nova.compute.resource_tracker [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 500.169396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.839s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.169396] env[61907]: DEBUG nova.service [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Creating RPC server for service compute {{(pid=61907) start /opt/stack/nova/nova/service.py:186}} [ 500.181065] env[61907]: DEBUG nova.service [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] Join ServiceGroup membership for this service compute {{(pid=61907) start /opt/stack/nova/nova/service.py:203}} [ 500.181065] env[61907]: DEBUG nova.servicegroup.drivers.db [None req-b2841671-6eaa-4729-bc9b-9075b63f6d85 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61907) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 538.295837] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquiring lock "ddd5ca18-61da-468f-a76c-8f939352aeda" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.297632] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "ddd5ca18-61da-468f-a76c-8f939352aeda" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.800982] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "7683c1aa-eec9-47ef-b775-0567f978da64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.801323] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "7683c1aa-eec9-47ef-b775-0567f978da64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.802623] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.308702] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 539.366732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.366732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.371620] env[61907]: INFO nova.compute.claims [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.763658] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquiring lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.763853] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.844933] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.268634] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 540.474605] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f606480c-fc1a-4b72-83e1-e8ceede95674 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.483664] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70a3376-7e3c-4243-988e-6cc9e0002777 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.518509] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44096614-c30f-4ba0-8acd-62dc2c6d658c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.527128] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625f3906-de9e-42ea-8e23-2a7d1469f149 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.540359] env[61907]: DEBUG nova.compute.provider_tree [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.792781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.043909] env[61907]: DEBUG nova.scheduler.client.report [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 541.549259] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.184s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.550808] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 541.554961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.710s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.555982] env[61907]: INFO nova.compute.claims [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.062416] env[61907]: DEBUG nova.compute.utils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.067212] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 542.067519] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 542.568497] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 542.654323] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452665f6-51ca-4e20-be43-c3ecfa66505a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.664290] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fc5e1f8-c8b0-4060-bf66-370799ee6136 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.706287] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bee1718-062c-48f7-9898-9da527cf2ebf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.718800] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc48c16-c810-4f80-9709-77a85efc0d7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.733398] env[61907]: DEBUG nova.compute.provider_tree [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 542.905892] env[61907]: DEBUG nova.policy [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ff2d0dc930f44beaabbfe0a0d5ab7e21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec16babab7de43aea04486771c9c131c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 543.240222] env[61907]: DEBUG nova.scheduler.client.report [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 543.579793] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 543.614158] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.615061] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.615061] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.615061] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.615061] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.615061] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.616863] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.617052] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.617598] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.617684] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.617802] env[61907]: DEBUG nova.virt.hardware [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.618817] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5f1401-ef06-4b3b-a50b-06ab5dd691bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.633167] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbab1590-4967-4f53-8476-099aefd86cfb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.648389] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dad4aa8-8901-433d-b2f7-8ffe01a0cbcd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.749206] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.195s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 543.750306] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 543.755463] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.961s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.757077] env[61907]: INFO nova.compute.claims [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.255527] env[61907]: DEBUG nova.compute.utils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.257094] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 544.262607] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 544.385891] env[61907]: DEBUG nova.policy [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10d6100ad10f4e6e93914e3cfa923aa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75a29d8205814a0b8ae2d5a45705312f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.616709] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "a3768960-4967-4975-9760-bc6ceb0b3e64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.617059] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "a3768960-4967-4975-9760-bc6ceb0b3e64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.673303] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Successfully created port: 026d734b-4942-472c-b0c8-402b289332dc {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.764761] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 544.882624] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647b4862-92d1-456c-8efb-0df724a49ce6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.889861] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf092aa9-fd9a-4b8b-be0e-07c6ed62b1a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.930989] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cc229f-4227-47e0-a4c6-2aae02e76c11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.939078] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d46b384-817d-4200-9265-8b50c2363060 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.953803] env[61907]: DEBUG nova.compute.provider_tree [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.120612] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 545.339823] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Successfully created port: 93fc454b-c9e0-4cd7-84ff-02c623c89eef {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.457661] env[61907]: DEBUG nova.scheduler.client.report [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 545.654573] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.775221] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 545.802875] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 545.805202] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 545.805202] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 545.805202] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 545.805202] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 545.805202] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 545.805357] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 545.805864] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 545.806256] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 545.806575] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 545.808637] env[61907]: DEBUG nova.virt.hardware [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 545.808637] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7b8de6-8cac-44f9-b587-e255db9be5ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.817880] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a12b6e-024b-4fbf-a9d0-2cfc2e81d9c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.966409] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.213s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.967128] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 545.972559] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.318s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.975414] env[61907]: INFO nova.compute.claims [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.479474] env[61907]: DEBUG nova.compute.utils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 546.485252] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 546.485252] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 546.697101] env[61907]: DEBUG nova.policy [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e294165a20a4b6f95c9a92d373b5285', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5efc878e429e4aa184c1cc2f122830d6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 546.983734] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 547.098771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d2b472-d51f-4bf5-968e-7573b25d6965 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.113016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9aca8e8-264e-4d27-994b-dd3457a9a7d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.148947] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c61ccf-c43b-4373-8a08-f15579df6742 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.157961] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e223d7d6-01bf-4ec7-a530-135f2a9cffc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.173916] env[61907]: DEBUG nova.compute.provider_tree [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.453468] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquiring lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.453781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.677494] env[61907]: DEBUG nova.scheduler.client.report [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 547.957366] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 547.999223] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 548.051936] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.053969] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.054227] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.054466] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.054560] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.054732] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.054912] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.055076] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.055255] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.055413] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.055577] env[61907]: DEBUG nova.virt.hardware [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.056540] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c381de5-dd62-4797-a9ca-9a607ae77341 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.066637] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c721ae03-02e3-4d12-8d80-8016bafb405e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.122026] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Successfully created port: ae7efd37-80d8-43d3-a2d3-d60ccdd77a74 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.183521] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.210s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.183521] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 548.490789] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.491057] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.492630] env[61907]: INFO nova.compute.claims [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.693022] env[61907]: DEBUG nova.compute.utils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 548.693022] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 548.693022] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 548.902404] env[61907]: DEBUG nova.policy [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 549.161023] env[61907]: ERROR nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 549.161023] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.161023] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.161023] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.161023] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.161023] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.161023] env[61907]: ERROR nova.compute.manager raise self.value [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.161023] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.161023] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.161023] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.162086] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.162086] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.162086] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 549.162086] env[61907]: ERROR nova.compute.manager [ 549.162086] env[61907]: Traceback (most recent call last): [ 549.162086] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.162086] env[61907]: listener.cb(fileno) [ 549.162086] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.162086] env[61907]: result = function(*args, **kwargs) [ 549.162086] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.162086] env[61907]: return func(*args, **kwargs) [ 549.162086] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.162086] env[61907]: raise e [ 549.162086] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.162086] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 549.162086] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.162086] env[61907]: created_port_ids = self._update_ports_for_instance( [ 549.162086] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.162086] env[61907]: with excutils.save_and_reraise_exception(): [ 549.162086] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.162086] env[61907]: self.force_reraise() [ 549.162086] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.162086] env[61907]: raise self.value [ 549.162086] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.162086] env[61907]: updated_port = self._update_port( [ 549.162086] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.162086] env[61907]: _ensure_no_port_binding_failure(port) [ 549.162086] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.162086] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.165446] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 549.165446] env[61907]: Removing descriptor: 16 [ 549.165446] env[61907]: ERROR nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Traceback (most recent call last): [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] yield resources [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.driver.spawn(context, instance, image_meta, [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.165446] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] vm_ref = self.build_virtual_machine(instance, [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] for vif in network_info: [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self._sync_wrapper(fn, *args, **kwargs) [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.wait() [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self[:] = self._gt.wait() [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self._exit_event.wait() [ 549.165742] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] result = hub.switch() [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self.greenlet.switch() [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] result = function(*args, **kwargs) [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return func(*args, **kwargs) [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise e [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] nwinfo = self.network_api.allocate_for_instance( [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.166103] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] created_port_ids = self._update_ports_for_instance( [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] with excutils.save_and_reraise_exception(): [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.force_reraise() [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise self.value [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] updated_port = self._update_port( [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] _ensure_no_port_binding_failure(port) [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.166431] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise exception.PortBindingFailed(port_id=port['id']) [ 549.167086] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 549.167086] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] [ 549.167086] env[61907]: INFO nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Terminating instance [ 549.202481] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 549.363432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquiring lock "a58bebcc-73e2-4752-befb-91083baacabf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.363432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "a58bebcc-73e2-4752-befb-91083baacabf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.617364] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6345bdc3-2eb3-4bb0-ab0c-1bb529b4ca2f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.626838] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eb35c33-1e1a-436e-a502-c5f3e057489a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.661714] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a861acef-18a9-4ad8-88ab-a54a2039af2d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.665592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.666766] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquired lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.666766] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.674131] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292c22f7-d9bd-44d2-a62b-a3881e6a3301 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.687670] env[61907]: DEBUG nova.compute.provider_tree [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.744080] env[61907]: ERROR nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 549.744080] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.744080] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.744080] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.744080] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.744080] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.744080] env[61907]: ERROR nova.compute.manager raise self.value [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.744080] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.744080] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.744080] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.744515] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.744515] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.744515] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 549.744515] env[61907]: ERROR nova.compute.manager [ 549.744515] env[61907]: Traceback (most recent call last): [ 549.744515] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.744515] env[61907]: listener.cb(fileno) [ 549.744515] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.744515] env[61907]: result = function(*args, **kwargs) [ 549.744515] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.744515] env[61907]: return func(*args, **kwargs) [ 549.744515] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.744515] env[61907]: raise e [ 549.744515] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.744515] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 549.744515] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.744515] env[61907]: created_port_ids = self._update_ports_for_instance( [ 549.744515] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.744515] env[61907]: with excutils.save_and_reraise_exception(): [ 549.744515] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.744515] env[61907]: self.force_reraise() [ 549.744515] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.744515] env[61907]: raise self.value [ 549.744515] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.744515] env[61907]: updated_port = self._update_port( [ 549.744515] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.744515] env[61907]: _ensure_no_port_binding_failure(port) [ 549.744515] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.744515] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.745263] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 549.745263] env[61907]: Removing descriptor: 15 [ 549.746085] env[61907]: ERROR nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Traceback (most recent call last): [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] yield resources [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.driver.spawn(context, instance, image_meta, [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] vm_ref = self.build_virtual_machine(instance, [ 549.746085] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] for vif in network_info: [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self._sync_wrapper(fn, *args, **kwargs) [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.wait() [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self[:] = self._gt.wait() [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self._exit_event.wait() [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.746382] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] result = hub.switch() [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self.greenlet.switch() [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] result = function(*args, **kwargs) [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return func(*args, **kwargs) [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise e [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] nwinfo = self.network_api.allocate_for_instance( [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] created_port_ids = self._update_ports_for_instance( [ 549.746681] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] with excutils.save_and_reraise_exception(): [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.force_reraise() [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise self.value [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] updated_port = self._update_port( [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] _ensure_no_port_binding_failure(port) [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise exception.PortBindingFailed(port_id=port['id']) [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 549.746990] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] [ 549.747400] env[61907]: INFO nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Terminating instance [ 549.867230] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 549.945677] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Successfully created port: ed13b448-e776-41b5-a2d8-77507d4d5ddd {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 550.182588] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.182588] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.182923] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 550.182923] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 550.190797] env[61907]: DEBUG nova.scheduler.client.report [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 550.213628] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 550.241370] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.248321] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 550.249880] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 550.249880] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 550.249880] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 550.249880] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 550.249880] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 550.250458] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 550.251270] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 550.251270] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 550.251401] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 550.252015] env[61907]: DEBUG nova.virt.hardware [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 550.253313] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a63b98d-ff90-4a6e-b1eb-2c844726e052 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.256596] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquiring lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.256753] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquired lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.256917] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.264233] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89438fb3-3d65-43cc-b8a7-21fe612a99cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.371696] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.401196] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.689870] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 550.691061] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 550.691265] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 550.691454] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 550.692486] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 550.692628] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Didn't find any instances for network info cache update. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 550.692844] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.693116] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.693313] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.693501] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.693688] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.693864] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_power_states {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 550.699978] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.208s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.700095] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 550.702344] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.301s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.704112] env[61907]: INFO nova.compute.claims [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.785869] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.876832] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Releasing lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.877411] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 550.877622] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.877882] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e05d70d4-1418-47ae-ba57-d5eff7053be7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.892711] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b2dead-7c50-49c4-a0b1-0c3be645d70c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.914865] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.921078] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7683c1aa-eec9-47ef-b775-0567f978da64 could not be found. [ 550.921572] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.921675] env[61907]: INFO nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.921932] env[61907]: DEBUG oslo.service.loopingcall [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.922374] env[61907]: DEBUG nova.compute.manager [-] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 550.922474] env[61907]: DEBUG nova.network.neutron [-] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.959765] env[61907]: DEBUG nova.network.neutron [-] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.165668] env[61907]: DEBUG nova.compute.manager [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Received event network-changed-93fc454b-c9e0-4cd7-84ff-02c623c89eef {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 551.165869] env[61907]: DEBUG nova.compute.manager [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Refreshing instance network info cache due to event network-changed-93fc454b-c9e0-4cd7-84ff-02c623c89eef. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 551.166108] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] Acquiring lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.166255] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] Acquired lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.166411] env[61907]: DEBUG nova.network.neutron [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Refreshing network info cache for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 551.198195] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Getting list of instances from cluster (obj){ [ 551.198195] env[61907]: value = "domain-c8" [ 551.198195] env[61907]: _type = "ClusterComputeResource" [ 551.198195] env[61907]: } {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 551.199479] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e87260-da7c-47b4-84d9-e1118e55315d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.216810] env[61907]: DEBUG nova.compute.utils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.220945] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Got total of 0 instances {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 551.220945] env[61907]: WARNING nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] While synchronizing instance power states, found 5 instances in the database and 0 instances on the hypervisor. [ 551.220945] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid ddd5ca18-61da-468f-a76c-8f939352aeda {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 551.220945] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid 7683c1aa-eec9-47ef-b775-0567f978da64 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 551.221556] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid fb4d09e4-6d60-41ed-862c-2bd8eca1e303 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 551.221556] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid a3768960-4967-4975-9760-bc6ceb0b3e64 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 551.221699] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid eb3a4f2a-8667-4d05-8606-a2dcb93c94a2 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 551.222471] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 551.223847] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 551.225017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "ddd5ca18-61da-468f-a76c-8f939352aeda" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.225392] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "7683c1aa-eec9-47ef-b775-0567f978da64" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.225663] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.225999] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "a3768960-4967-4975-9760-bc6ceb0b3e64" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.226263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.226494] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.226754] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 551.227989] env[61907]: ERROR nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 551.227989] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.227989] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.227989] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.227989] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.227989] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.227989] env[61907]: ERROR nova.compute.manager raise self.value [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.227989] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.227989] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.227989] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.228485] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.228485] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.228485] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 551.228485] env[61907]: ERROR nova.compute.manager [ 551.228485] env[61907]: Traceback (most recent call last): [ 551.228625] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.228625] env[61907]: listener.cb(fileno) [ 551.228625] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.228625] env[61907]: result = function(*args, **kwargs) [ 551.228625] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.228625] env[61907]: return func(*args, **kwargs) [ 551.228625] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.228625] env[61907]: raise e [ 551.228625] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.228625] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 551.228625] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.228625] env[61907]: created_port_ids = self._update_ports_for_instance( [ 551.228625] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.228625] env[61907]: with excutils.save_and_reraise_exception(): [ 551.228625] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.228625] env[61907]: self.force_reraise() [ 551.228625] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.228625] env[61907]: raise self.value [ 551.228625] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.228625] env[61907]: updated_port = self._update_port( [ 551.228625] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.228625] env[61907]: _ensure_no_port_binding_failure(port) [ 551.228625] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.228625] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.228625] env[61907]: nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 551.228625] env[61907]: Removing descriptor: 17 [ 551.229470] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 551.231361] env[61907]: ERROR nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Traceback (most recent call last): [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] yield resources [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.driver.spawn(context, instance, image_meta, [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] vm_ref = self.build_virtual_machine(instance, [ 551.231361] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] for vif in network_info: [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self._sync_wrapper(fn, *args, **kwargs) [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.wait() [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self[:] = self._gt.wait() [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self._exit_event.wait() [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.231767] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] result = hub.switch() [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self.greenlet.switch() [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] result = function(*args, **kwargs) [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return func(*args, **kwargs) [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise e [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] nwinfo = self.network_api.allocate_for_instance( [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] created_port_ids = self._update_ports_for_instance( [ 551.232185] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] with excutils.save_and_reraise_exception(): [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.force_reraise() [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise self.value [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] updated_port = self._update_port( [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] _ensure_no_port_binding_failure(port) [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise exception.PortBindingFailed(port_id=port['id']) [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 551.232546] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] [ 551.233024] env[61907]: INFO nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Terminating instance [ 551.291899] env[61907]: DEBUG nova.policy [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e582fd6f6c3944dfabe41439166e0d19', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fb5c3d12b927433fa7a1b7e8e5be247c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 551.423557] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Releasing lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.426018] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 551.426018] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 551.426018] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b1e49df8-76dd-4e3a-855c-1d3db726aa52 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.436222] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cd014e-2f2e-49c7-9c62-e40e406184b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.461305] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ddd5ca18-61da-468f-a76c-8f939352aeda could not be found. [ 551.461580] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 551.461900] env[61907]: INFO nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Took 0.04 seconds to destroy the instance on the hypervisor. [ 551.462465] env[61907]: DEBUG oslo.service.loopingcall [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.462725] env[61907]: DEBUG nova.network.neutron [-] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.464499] env[61907]: DEBUG nova.compute.manager [-] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 551.464663] env[61907]: DEBUG nova.network.neutron [-] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 551.468699] env[61907]: DEBUG nova.compute.manager [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Received event network-changed-026d734b-4942-472c-b0c8-402b289332dc {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 551.468900] env[61907]: DEBUG nova.compute.manager [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Refreshing instance network info cache due to event network-changed-026d734b-4942-472c-b0c8-402b289332dc. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 551.469136] env[61907]: DEBUG oslo_concurrency.lockutils [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] Acquiring lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.470901] env[61907]: DEBUG oslo_concurrency.lockutils [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] Acquired lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.470901] env[61907]: DEBUG nova.network.neutron [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Refreshing network info cache for port 026d734b-4942-472c-b0c8-402b289332dc {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 551.505731] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquiring lock "c45186da-90cb-4551-a7be-4deeae587967" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.505731] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "c45186da-90cb-4551-a7be-4deeae587967" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.515681] env[61907]: ERROR nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 551.515681] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.515681] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.515681] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.515681] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.515681] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.515681] env[61907]: ERROR nova.compute.manager raise self.value [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.515681] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.515681] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.515681] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.516177] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.516177] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.516177] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 551.516177] env[61907]: ERROR nova.compute.manager [ 551.516177] env[61907]: Traceback (most recent call last): [ 551.516177] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.516177] env[61907]: listener.cb(fileno) [ 551.516177] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.516177] env[61907]: result = function(*args, **kwargs) [ 551.516177] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.516177] env[61907]: return func(*args, **kwargs) [ 551.516177] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.516177] env[61907]: raise e [ 551.516177] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.516177] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 551.516177] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.516177] env[61907]: created_port_ids = self._update_ports_for_instance( [ 551.516177] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.516177] env[61907]: with excutils.save_and_reraise_exception(): [ 551.516177] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.516177] env[61907]: self.force_reraise() [ 551.516177] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.516177] env[61907]: raise self.value [ 551.516177] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.516177] env[61907]: updated_port = self._update_port( [ 551.516177] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.516177] env[61907]: _ensure_no_port_binding_failure(port) [ 551.516177] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.516177] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.516907] env[61907]: nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 551.516907] env[61907]: Removing descriptor: 18 [ 551.516979] env[61907]: ERROR nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Traceback (most recent call last): [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] yield resources [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.driver.spawn(context, instance, image_meta, [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] vm_ref = self.build_virtual_machine(instance, [ 551.516979] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] for vif in network_info: [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self._sync_wrapper(fn, *args, **kwargs) [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.wait() [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self[:] = self._gt.wait() [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self._exit_event.wait() [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.517301] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] result = hub.switch() [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self.greenlet.switch() [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] result = function(*args, **kwargs) [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return func(*args, **kwargs) [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise e [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] nwinfo = self.network_api.allocate_for_instance( [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] created_port_ids = self._update_ports_for_instance( [ 551.517647] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] with excutils.save_and_reraise_exception(): [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.force_reraise() [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise self.value [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] updated_port = self._update_port( [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] _ensure_no_port_binding_failure(port) [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise exception.PortBindingFailed(port_id=port['id']) [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 551.519145] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] [ 551.519713] env[61907]: INFO nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Terminating instance [ 551.519713] env[61907]: DEBUG nova.network.neutron [-] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.692499] env[61907]: DEBUG nova.network.neutron [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.725622] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 551.739022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.739590] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquiring lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.739742] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquired lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.739909] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 551.768139] env[61907]: DEBUG nova.network.neutron [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.843311] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Successfully created port: 4aa89582-e057-45e2-ab6a-da8aca9f219c {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 551.885965] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2bb885-b21d-4146-bdaa-96ad13d14ade {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.894096] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2424d051-e10d-408c-ac96-6390ad9ad69c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.926360] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f819a666-d2ed-440d-a493-a6be9651bef6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.936739] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc173c47-6c4e-4e1f-acab-3d2a34a43fee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.953402] env[61907]: DEBUG nova.compute.provider_tree [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 551.969345] env[61907]: INFO nova.compute.manager [-] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Took 1.05 seconds to deallocate network for instance. [ 551.975827] env[61907]: DEBUG nova.compute.claims [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.976047] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.008284] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 552.014207] env[61907]: DEBUG nova.network.neutron [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.022405] env[61907]: DEBUG nova.network.neutron [-] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.026478] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.026643] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.026808] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 552.179028] env[61907]: DEBUG nova.network.neutron [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.270269] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.275098] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef55f4d6-5c9e-418e-8b16-5510f11e84be req-2579b5d5-4a8a-40e4-90c8-3765cae8c879 service nova] Releasing lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.365742] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.459031] env[61907]: DEBUG nova.scheduler.client.report [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 552.528801] env[61907]: INFO nova.compute.manager [-] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Took 1.06 seconds to deallocate network for instance. [ 552.530955] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.533310] env[61907]: DEBUG nova.compute.claims [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.533454] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.586766] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.683162] env[61907]: DEBUG oslo_concurrency.lockutils [req-df828bc1-a8c4-40bf-93c2-3eb25e8087e4 req-d6e6e293-624b-4197-b708-ab1040f5afc9 service nova] Releasing lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.738296] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 552.766629] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.766629] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.766866] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.766930] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.767100] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.767280] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.767454] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.767610] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.767827] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.767923] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.768350] env[61907]: DEBUG nova.virt.hardware [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.769248] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f149756-24d4-491c-8ec7-48935f63ecb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.777673] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd685214-53b7-49bc-8ae7-2f1e67bebe01 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.823074] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.868614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Releasing lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.869009] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 552.869215] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 552.869497] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed94a1a5-55ca-417c-a6c1-cdfe5cb1757a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.879371] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b604208-0701-4e02-9a35-6f3561a550ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.905447] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb4d09e4-6d60-41ed-862c-2bd8eca1e303 could not be found. [ 552.905726] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 552.905978] env[61907]: INFO nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Took 0.04 seconds to destroy the instance on the hypervisor. [ 552.906277] env[61907]: DEBUG oslo.service.loopingcall [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.907060] env[61907]: DEBUG nova.compute.manager [-] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 552.907656] env[61907]: DEBUG nova.network.neutron [-] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.933595] env[61907]: DEBUG nova.network.neutron [-] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.964687] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.965243] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 552.968357] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.229s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.968505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.968685] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 552.968974] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 0.993s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.972997] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b023f39-daf0-4671-aa1e-9e98f834bafa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.983549] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b71ba9f-731c-4e7f-9129-4f36a9561b4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.991254] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquiring lock "dc04ac15-a97a-4248-b578-8f9046ea7e8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.991417] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "dc04ac15-a97a-4248-b578-8f9046ea7e8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.006025] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f51147-1ae2-4576-84e4-0e115b97e215 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.017169] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d91d64-3835-4fb3-915a-d1ec1375c2e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.048187] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181535MB free_disk=149GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 553.049042] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.328834] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.329960] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 553.329960] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 553.329960] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4fe52a13-fb71-4136-bbe1-1085f6057b57 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.342928] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560fc73b-ba54-47c4-b37b-a3f4655f341b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.369938] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a3768960-4967-4975-9760-bc6ceb0b3e64 could not be found. [ 553.370319] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 553.370665] env[61907]: INFO nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Took 0.04 seconds to destroy the instance on the hypervisor. [ 553.371163] env[61907]: DEBUG oslo.service.loopingcall [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.371488] env[61907]: DEBUG nova.compute.manager [-] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 553.371678] env[61907]: DEBUG nova.network.neutron [-] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 553.391762] env[61907]: DEBUG nova.network.neutron [-] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 553.437233] env[61907]: DEBUG nova.network.neutron [-] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.471580] env[61907]: DEBUG nova.compute.utils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.472784] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 553.473514] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.496710] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 553.605786] env[61907]: DEBUG nova.policy [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '608a09df66ba4b8e9af79adcea761170', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bd37a3beb6954356be3f6380c62a4d73', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 553.620117] env[61907]: ERROR nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 553.620117] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.620117] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.620117] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.620117] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.620117] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.620117] env[61907]: ERROR nova.compute.manager raise self.value [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.620117] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 553.620117] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.620117] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 553.620522] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.620522] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 553.620522] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 553.620522] env[61907]: ERROR nova.compute.manager [ 553.620522] env[61907]: Traceback (most recent call last): [ 553.620522] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 553.620522] env[61907]: listener.cb(fileno) [ 553.620522] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.620522] env[61907]: result = function(*args, **kwargs) [ 553.620522] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.620522] env[61907]: return func(*args, **kwargs) [ 553.620522] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.620522] env[61907]: raise e [ 553.620522] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.620522] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 553.620522] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.620522] env[61907]: created_port_ids = self._update_ports_for_instance( [ 553.620522] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.620522] env[61907]: with excutils.save_and_reraise_exception(): [ 553.620522] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.620522] env[61907]: self.force_reraise() [ 553.620522] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.620522] env[61907]: raise self.value [ 553.620522] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.620522] env[61907]: updated_port = self._update_port( [ 553.620522] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.620522] env[61907]: _ensure_no_port_binding_failure(port) [ 553.620522] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.620522] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 553.621261] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 553.621261] env[61907]: Removing descriptor: 15 [ 553.621261] env[61907]: ERROR nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Traceback (most recent call last): [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] yield resources [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.driver.spawn(context, instance, image_meta, [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.621261] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] vm_ref = self.build_virtual_machine(instance, [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] for vif in network_info: [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self._sync_wrapper(fn, *args, **kwargs) [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.wait() [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self[:] = self._gt.wait() [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self._exit_event.wait() [ 553.621638] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] result = hub.switch() [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self.greenlet.switch() [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] result = function(*args, **kwargs) [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return func(*args, **kwargs) [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise e [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] nwinfo = self.network_api.allocate_for_instance( [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 553.622017] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] created_port_ids = self._update_ports_for_instance( [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] with excutils.save_and_reraise_exception(): [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.force_reraise() [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise self.value [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] updated_port = self._update_port( [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] _ensure_no_port_binding_failure(port) [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.622444] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise exception.PortBindingFailed(port_id=port['id']) [ 553.622759] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 553.622759] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] [ 553.622759] env[61907]: INFO nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Terminating instance [ 553.639570] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebda3028-f6fc-4b95-890f-2d60632e06cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.647840] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95714cbb-2b17-4c2c-b04a-a858c9ecfffe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.686097] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78e5b40-614d-41d4-9b8b-5033f661f193 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.695823] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98cb685-a438-4d03-9d12-3504740ae06b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.712273] env[61907]: DEBUG nova.compute.provider_tree [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.896499] env[61907]: DEBUG nova.network.neutron [-] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.940696] env[61907]: INFO nova.compute.manager [-] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Took 1.03 seconds to deallocate network for instance. [ 553.943694] env[61907]: DEBUG nova.compute.claims [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 553.943694] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.981507] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 554.026330] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.126760] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquiring lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.126984] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquired lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.128025] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.186683] env[61907]: DEBUG nova.compute.manager [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Received event network-changed-ed13b448-e776-41b5-a2d8-77507d4d5ddd {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 554.187032] env[61907]: DEBUG nova.compute.manager [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Refreshing instance network info cache due to event network-changed-ed13b448-e776-41b5-a2d8-77507d4d5ddd. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 554.188622] env[61907]: DEBUG oslo_concurrency.lockutils [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] Acquiring lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.188622] env[61907]: DEBUG oslo_concurrency.lockutils [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] Acquired lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.188622] env[61907]: DEBUG nova.network.neutron [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Refreshing network info cache for port ed13b448-e776-41b5-a2d8-77507d4d5ddd {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.193226] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Successfully created port: a2cc7c85-6d99-4eb1-85aa-023d902efaef {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 554.215241] env[61907]: DEBUG nova.scheduler.client.report [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 554.386515] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Received event network-vif-deleted-026d734b-4942-472c-b0c8-402b289332dc {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 554.386747] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Received event network-changed-ae7efd37-80d8-43d3-a2d3-d60ccdd77a74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 554.386928] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Refreshing instance network info cache due to event network-changed-ae7efd37-80d8-43d3-a2d3-d60ccdd77a74. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 554.390429] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Acquiring lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.390636] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Acquired lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.390810] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Refreshing network info cache for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 554.398793] env[61907]: INFO nova.compute.manager [-] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Took 1.03 seconds to deallocate network for instance. [ 554.402269] env[61907]: DEBUG nova.compute.claims [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 554.402458] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.666771] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.699592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquiring lock "4db46e0e-1aaa-4168-8365-6dab93eee759" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.699879] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "4db46e0e-1aaa-4168-8365-6dab93eee759" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.724835] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.756s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.725591] env[61907]: ERROR nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Traceback (most recent call last): [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.driver.spawn(context, instance, image_meta, [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] vm_ref = self.build_virtual_machine(instance, [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.725591] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] for vif in network_info: [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self._sync_wrapper(fn, *args, **kwargs) [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.wait() [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self[:] = self._gt.wait() [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self._exit_event.wait() [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] result = hub.switch() [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.726087] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return self.greenlet.switch() [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] result = function(*args, **kwargs) [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] return func(*args, **kwargs) [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise e [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] nwinfo = self.network_api.allocate_for_instance( [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] created_port_ids = self._update_ports_for_instance( [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] with excutils.save_and_reraise_exception(): [ 554.726578] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] self.force_reraise() [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise self.value [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] updated_port = self._update_port( [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] _ensure_no_port_binding_failure(port) [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] raise exception.PortBindingFailed(port_id=port['id']) [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] nova.exception.PortBindingFailed: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. [ 554.729910] env[61907]: ERROR nova.compute.manager [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] [ 554.730299] env[61907]: DEBUG nova.compute.utils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 554.730299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.197s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.730299] env[61907]: INFO nova.compute.claims [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.737811] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Build of instance 7683c1aa-eec9-47ef-b775-0567f978da64 was re-scheduled: Binding failed for port 93fc454b-c9e0-4cd7-84ff-02c623c89eef, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 554.738373] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 554.738601] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.739013] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquired lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.739013] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 554.743104] env[61907]: DEBUG nova.network.neutron [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.867740] env[61907]: DEBUG nova.network.neutron [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.911435] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 554.972022] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.991185] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 555.007293] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.017664] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.017965] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.018077] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.018413] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.018578] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.018729] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.018926] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.019159] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.019573] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.019672] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.019855] env[61907]: DEBUG nova.virt.hardware [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.021515] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e13c5f0e-523b-4d01-a42d-625811dfa5d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.030642] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225ed9a4-167c-413c-87f9-fdba1d51ed30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.206097] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 555.278129] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.372405] env[61907]: DEBUG oslo_concurrency.lockutils [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] Releasing lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.372405] env[61907]: DEBUG nova.compute.manager [req-4de9390f-d9fe-4d5f-bc10-daa74dd71382 req-d05bf8d1-499d-44b7-b9fb-803b28241c76 service nova] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Received event network-vif-deleted-ed13b448-e776-41b5-a2d8-77507d4d5ddd {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 555.403154] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.475449] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Releasing lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.475731] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Received event network-vif-deleted-ae7efd37-80d8-43d3-a2d3-d60ccdd77a74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 555.475892] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Received event network-changed-4aa89582-e057-45e2-ab6a-da8aca9f219c {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 555.476088] env[61907]: DEBUG nova.compute.manager [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Refreshing instance network info cache due to event network-changed-4aa89582-e057-45e2-ab6a-da8aca9f219c. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 555.476293] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Acquiring lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.510775] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Releasing lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.512125] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 555.512810] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 555.512899] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Acquired lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.513117] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Refreshing network info cache for port 4aa89582-e057-45e2-ab6a-da8aca9f219c {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 555.515376] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fecabd4a-f395-41ff-b091-176551e39d1c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.525892] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d1d3ee-031e-43a5-89aa-4e6ead22dd0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.558403] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eb3a4f2a-8667-4d05-8606-a2dcb93c94a2 could not be found. [ 555.558698] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 555.558936] env[61907]: INFO nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Took 0.05 seconds to destroy the instance on the hypervisor. [ 555.559561] env[61907]: DEBUG oslo.service.loopingcall [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.559561] env[61907]: DEBUG nova.compute.manager [-] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 555.559833] env[61907]: DEBUG nova.network.neutron [-] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.584137] env[61907]: DEBUG nova.network.neutron [-] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 555.741832] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.883783] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d92ab9-e1dc-40bf-a3ee-d7e0c402aa24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.893091] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99411721-c439-4c09-8618-46ab367ce7f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.925063] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Releasing lock "refresh_cache-7683c1aa-eec9-47ef-b775-0567f978da64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.925354] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 555.925558] env[61907]: DEBUG nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 555.925761] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 555.928383] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7900b8b-50e6-4805-b1ec-f2c6e711ac65 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.935751] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcde7b2e-9003-4638-880b-219524cae0cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.951565] env[61907]: DEBUG nova.compute.provider_tree [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.972851] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.063658] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.086441] env[61907]: DEBUG nova.network.neutron [-] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.457105] env[61907]: DEBUG nova.scheduler.client.report [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 556.461466] env[61907]: DEBUG nova.network.neutron [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.476759] env[61907]: DEBUG nova.network.neutron [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.588940] env[61907]: INFO nova.compute.manager [-] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Took 1.03 seconds to deallocate network for instance. [ 556.591887] env[61907]: DEBUG nova.compute.claims [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 556.592377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.964493] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.966428] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 556.968184] env[61907]: DEBUG oslo_concurrency.lockutils [req-29a207dc-7f0e-45bf-8632-7e350778da49 req-429ec302-3ceb-4207-80c4-d8a8d2aec0c9 service nova] Releasing lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.968423] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.435s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.985809] env[61907]: INFO nova.compute.manager [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] Took 1.06 seconds to deallocate network for instance. [ 557.473197] env[61907]: DEBUG nova.compute.utils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.478589] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 557.478762] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 557.511539] env[61907]: DEBUG nova.compute.manager [req-52d91b98-6dd2-442b-a8b7-542bf829e16f req-9f75d80b-d7b2-4c29-bb28-2d9a59ca7d00 service nova] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Received event network-vif-deleted-4aa89582-e057-45e2-ab6a-da8aca9f219c {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 557.532987] env[61907]: ERROR nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 557.532987] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 557.532987] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.532987] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.532987] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.532987] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.532987] env[61907]: ERROR nova.compute.manager raise self.value [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.532987] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.532987] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.532987] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.533444] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.533444] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.533444] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 557.533444] env[61907]: ERROR nova.compute.manager [ 557.533444] env[61907]: Traceback (most recent call last): [ 557.533444] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.533444] env[61907]: listener.cb(fileno) [ 557.533444] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.533444] env[61907]: result = function(*args, **kwargs) [ 557.533444] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.533444] env[61907]: return func(*args, **kwargs) [ 557.533444] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 557.533444] env[61907]: raise e [ 557.533444] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 557.533444] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 557.533444] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.533444] env[61907]: created_port_ids = self._update_ports_for_instance( [ 557.533444] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.533444] env[61907]: with excutils.save_and_reraise_exception(): [ 557.533444] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.533444] env[61907]: self.force_reraise() [ 557.533444] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.533444] env[61907]: raise self.value [ 557.533444] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.533444] env[61907]: updated_port = self._update_port( [ 557.533444] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.533444] env[61907]: _ensure_no_port_binding_failure(port) [ 557.533444] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.533444] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.535466] env[61907]: nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 557.535466] env[61907]: Removing descriptor: 17 [ 557.535466] env[61907]: ERROR nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] Traceback (most recent call last): [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] yield resources [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.driver.spawn(context, instance, image_meta, [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.535466] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] vm_ref = self.build_virtual_machine(instance, [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] for vif in network_info: [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self._sync_wrapper(fn, *args, **kwargs) [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.wait() [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self[:] = self._gt.wait() [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self._exit_event.wait() [ 557.535854] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] result = hub.switch() [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self.greenlet.switch() [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] result = function(*args, **kwargs) [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return func(*args, **kwargs) [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise e [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] nwinfo = self.network_api.allocate_for_instance( [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 557.536233] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] created_port_ids = self._update_ports_for_instance( [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] with excutils.save_and_reraise_exception(): [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.force_reraise() [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise self.value [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] updated_port = self._update_port( [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] _ensure_no_port_binding_failure(port) [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.536621] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise exception.PortBindingFailed(port_id=port['id']) [ 557.536997] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 557.536997] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] [ 557.536997] env[61907]: INFO nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Terminating instance [ 557.647527] env[61907]: DEBUG nova.policy [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7eebf7ecea70448bb2bb1192768941f7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f20a860783d4ad9bd837022fd17bbfa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 557.717276] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc6f643-bf24-48a1-8407-c894be11c90a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.728079] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34461f7e-b0b5-4fdc-a154-9df4641c57c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.769781] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ad8de0-4e9d-48a8-bc3a-ea3853f6e493 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.778651] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142cda40-e73b-49a1-8310-ad5ac9529c50 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.798235] env[61907]: DEBUG nova.compute.provider_tree [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.981484] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 558.033863] env[61907]: INFO nova.scheduler.client.report [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Deleted allocations for instance 7683c1aa-eec9-47ef-b775-0567f978da64 [ 558.044233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquiring lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.044233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquired lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.044233] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.303991] env[61907]: DEBUG nova.scheduler.client.report [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 558.549409] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1d08d8ab-8232-44bf-8428-5f154a0f2cab tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "7683c1aa-eec9-47ef-b775-0567f978da64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.747s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.549828] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "7683c1aa-eec9-47ef-b775-0567f978da64" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 7.324s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.549828] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 7683c1aa-eec9-47ef-b775-0567f978da64] During sync_power_state the instance has a pending task (spawning). Skip. [ 558.550537] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "7683c1aa-eec9-47ef-b775-0567f978da64" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.601437] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 558.806885] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Successfully created port: e37bdea3-0728-41ac-86f4-95ff72e30057 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.809575] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.841s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.812485] env[61907]: ERROR nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Traceback (most recent call last): [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.driver.spawn(context, instance, image_meta, [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] vm_ref = self.build_virtual_machine(instance, [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.812485] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] for vif in network_info: [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self._sync_wrapper(fn, *args, **kwargs) [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.wait() [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self[:] = self._gt.wait() [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self._exit_event.wait() [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] result = hub.switch() [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.812972] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return self.greenlet.switch() [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] result = function(*args, **kwargs) [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] return func(*args, **kwargs) [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise e [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] nwinfo = self.network_api.allocate_for_instance( [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] created_port_ids = self._update_ports_for_instance( [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] with excutils.save_and_reraise_exception(): [ 558.813370] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] self.force_reraise() [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise self.value [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] updated_port = self._update_port( [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] _ensure_no_port_binding_failure(port) [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] raise exception.PortBindingFailed(port_id=port['id']) [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] nova.exception.PortBindingFailed: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. [ 558.814064] env[61907]: ERROR nova.compute.manager [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] [ 558.814388] env[61907]: DEBUG nova.compute.utils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 558.814388] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Build of instance ddd5ca18-61da-468f-a76c-8f939352aeda was re-scheduled: Binding failed for port 026d734b-4942-472c-b0c8-402b289332dc, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 558.814493] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 558.814894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquiring lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.815096] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Acquired lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.815536] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 558.816833] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.768s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.835071] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.990279] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 559.020981] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.021350] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.021481] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.021673] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.021809] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.022059] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.022610] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.022806] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.023014] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.023145] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.023328] env[61907]: DEBUG nova.virt.hardware [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.024250] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0306bfcd-b5be-404d-9c86-ae6c3bd1afaf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.033298] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d8049db-5444-4804-9cc7-9c65e5b9ca7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.340498] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Releasing lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.341873] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 559.341873] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.341873] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-59b19619-a233-4a51-961f-448ad884bf13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.354456] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-803f487e-c689-4c92-a1e3-958d56a54033 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.373451] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.380353] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a58bebcc-73e2-4752-befb-91083baacabf could not be found. [ 559.380535] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 559.380784] env[61907]: INFO nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.381328] env[61907]: DEBUG oslo.service.loopingcall [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.381328] env[61907]: DEBUG nova.compute.manager [-] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 559.381452] env[61907]: DEBUG nova.network.neutron [-] [instance: a58bebcc-73e2-4752-befb-91083baacabf] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 559.503149] env[61907]: DEBUG nova.network.neutron [-] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 559.828322] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.008463] env[61907]: DEBUG nova.network.neutron [-] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.332235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Releasing lock "refresh_cache-ddd5ca18-61da-468f-a76c-8f939352aeda" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.332635] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 560.332823] env[61907]: DEBUG nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 560.332984] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 560.372697] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ddd5ca18-61da-468f-a76c-8f939352aeda has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.372697] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance fb4d09e4-6d60-41ed-862c-2bd8eca1e303 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 560.372697] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a3768960-4967-4975-9760-bc6ceb0b3e64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 560.372923] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance eb3a4f2a-8667-4d05-8606-a2dcb93c94a2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 560.373154] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a58bebcc-73e2-4752-befb-91083baacabf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 560.373306] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c45186da-90cb-4551-a7be-4deeae587967 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 560.418846] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 560.490743] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquiring lock "71013ca4-f1d0-4110-8fad-91e619ff354b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.491028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "71013ca4-f1d0-4110-8fad-91e619ff354b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.514564] env[61907]: INFO nova.compute.manager [-] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Took 1.13 seconds to deallocate network for instance. [ 560.521403] env[61907]: DEBUG nova.compute.claims [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.521669] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.617109] env[61907]: DEBUG nova.compute.manager [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Received event network-changed-a2cc7c85-6d99-4eb1-85aa-023d902efaef {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 560.617331] env[61907]: DEBUG nova.compute.manager [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Refreshing instance network info cache due to event network-changed-a2cc7c85-6d99-4eb1-85aa-023d902efaef. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 560.617535] env[61907]: DEBUG oslo_concurrency.lockutils [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] Acquiring lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.617677] env[61907]: DEBUG oslo_concurrency.lockutils [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] Acquired lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.617831] env[61907]: DEBUG nova.network.neutron [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Refreshing network info cache for port a2cc7c85-6d99-4eb1-85aa-023d902efaef {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.878409] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance dc04ac15-a97a-4248-b578-8f9046ea7e8e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.923933] env[61907]: DEBUG nova.network.neutron [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.997424] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 561.186498] env[61907]: DEBUG nova.network.neutron [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.381326] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4db46e0e-1aaa-4168-8365-6dab93eee759 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.381663] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 561.381764] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 561.434342] env[61907]: INFO nova.compute.manager [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] Took 1.10 seconds to deallocate network for instance. [ 561.524041] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.554187] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daec7c18-f972-4ddd-b5af-9d6e3c1f9c47 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.562357] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e9c48c-4da3-40a7-9582-ec9bedf32cd4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.595917] env[61907]: DEBUG nova.network.neutron [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.597544] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0194d3ed-b32a-4dfb-94c3-2b9ffa9d753b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.605707] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6dc568d-55ff-4db8-9998-7ed6761af42a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.619594] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 562.102731] env[61907]: DEBUG oslo_concurrency.lockutils [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] Releasing lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.102731] env[61907]: DEBUG nova.compute.manager [req-d6938b46-e94e-4ba9-a130-4ada6adaf9c7 req-e85e736f-10d3-4e2f-9399-deb140d5563d service nova] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Received event network-vif-deleted-a2cc7c85-6d99-4eb1-85aa-023d902efaef {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 562.157603] env[61907]: ERROR nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [req-86c5d211-8e11-4966-9a17-93ac36eee92f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-86c5d211-8e11-4966-9a17-93ac36eee92f"}]} [ 562.195140] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 562.220808] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 562.221104] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 149, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 562.243811] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 562.275593] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 562.497489] env[61907]: INFO nova.scheduler.client.report [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Deleted allocations for instance ddd5ca18-61da-468f-a76c-8f939352aeda [ 562.545060] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d774322-5eb5-4aff-ae47-59295933fb42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.554203] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c776094-9eae-4e89-b7e5-683edb59c386 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.587486] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e9d10e-28b3-452f-82f3-f7055d6e3001 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.599927] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22a6419-65e0-4fb8-8181-eeb1c8bb0988 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.615169] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 563.022347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0de06863-2880-40e7-95f3-24ffeebc1b93 tempest-ServerDiagnosticsTest-246055975 tempest-ServerDiagnosticsTest-246055975-project-member] Lock "ddd5ca18-61da-468f-a76c-8f939352aeda" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.725s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.022347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "ddd5ca18-61da-468f-a76c-8f939352aeda" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.797s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.022347] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ddd5ca18-61da-468f-a76c-8f939352aeda] During sync_power_state the instance has a pending task (spawning). Skip. [ 563.022347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "ddd5ca18-61da-468f-a76c-8f939352aeda" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.168265] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 14 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 563.168955] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 14 to 15 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 563.168955] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 563.453765] env[61907]: ERROR nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 563.453765] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.453765] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.453765] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.453765] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.453765] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.453765] env[61907]: ERROR nova.compute.manager raise self.value [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.453765] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 563.453765] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.453765] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 563.454708] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.454708] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 563.454708] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 563.454708] env[61907]: ERROR nova.compute.manager [ 563.454708] env[61907]: Traceback (most recent call last): [ 563.454708] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 563.454708] env[61907]: listener.cb(fileno) [ 563.454708] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.454708] env[61907]: result = function(*args, **kwargs) [ 563.454708] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.454708] env[61907]: return func(*args, **kwargs) [ 563.454708] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 563.454708] env[61907]: raise e [ 563.454708] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.454708] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 563.454708] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.454708] env[61907]: created_port_ids = self._update_ports_for_instance( [ 563.454708] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.454708] env[61907]: with excutils.save_and_reraise_exception(): [ 563.454708] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.454708] env[61907]: self.force_reraise() [ 563.454708] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.454708] env[61907]: raise self.value [ 563.454708] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.454708] env[61907]: updated_port = self._update_port( [ 563.454708] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.454708] env[61907]: _ensure_no_port_binding_failure(port) [ 563.454708] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.454708] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 563.455458] env[61907]: nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 563.455458] env[61907]: Removing descriptor: 15 [ 563.455458] env[61907]: ERROR nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] Traceback (most recent call last): [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] yield resources [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.driver.spawn(context, instance, image_meta, [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self._vmops.spawn(context, instance, image_meta, injected_files, [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 563.455458] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] vm_ref = self.build_virtual_machine(instance, [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] vif_infos = vmwarevif.get_vif_info(self._session, [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] for vif in network_info: [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self._sync_wrapper(fn, *args, **kwargs) [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.wait() [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self[:] = self._gt.wait() [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self._exit_event.wait() [ 563.455782] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] result = hub.switch() [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self.greenlet.switch() [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] result = function(*args, **kwargs) [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return func(*args, **kwargs) [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise e [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] nwinfo = self.network_api.allocate_for_instance( [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 563.459305] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] created_port_ids = self._update_ports_for_instance( [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] with excutils.save_and_reraise_exception(): [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.force_reraise() [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise self.value [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] updated_port = self._update_port( [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] _ensure_no_port_binding_failure(port) [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 563.459645] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise exception.PortBindingFailed(port_id=port['id']) [ 563.459951] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 563.459951] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] [ 563.459951] env[61907]: INFO nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Terminating instance [ 563.676330] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 563.676568] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.860s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.676824] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.733s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.679766] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 563.680429] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Getting list of instances from cluster (obj){ [ 563.680429] env[61907]: value = "domain-c8" [ 563.680429] env[61907]: _type = "ClusterComputeResource" [ 563.680429] env[61907]: } {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 563.685722] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3f4654-16a1-4989-b0da-6d70fb6c7cf5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.698930] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Got total of 0 instances {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 563.913117] env[61907]: DEBUG nova.compute.manager [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Received event network-changed-e37bdea3-0728-41ac-86f4-95ff72e30057 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 563.913117] env[61907]: DEBUG nova.compute.manager [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Refreshing instance network info cache due to event network-changed-e37bdea3-0728-41ac-86f4-95ff72e30057. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 563.913117] env[61907]: DEBUG oslo_concurrency.lockutils [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] Acquiring lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 563.913117] env[61907]: DEBUG oslo_concurrency.lockutils [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] Acquired lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 563.913117] env[61907]: DEBUG nova.network.neutron [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Refreshing network info cache for port e37bdea3-0728-41ac-86f4-95ff72e30057 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 563.961976] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquiring lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 564.360122] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259dc9b3-550e-4442-9d26-0a25f049e0d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.370962] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b80ae9ff-f5ea-4abf-9865-c85f94f27708 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.428849] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85161116-4ad2-4729-8a19-2cbb3a0e71d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.439326] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c24aad0-ea63-4dc5-801a-de6a8993051c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.455311] env[61907]: DEBUG nova.compute.provider_tree [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 564.481062] env[61907]: DEBUG nova.network.neutron [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 564.684208] env[61907]: DEBUG nova.network.neutron [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.958789] env[61907]: DEBUG nova.scheduler.client.report [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 565.026583] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquiring lock "f66cffbb-1af4-4675-ae4c-ea6a034217c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.027166] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "f66cffbb-1af4-4675-ae4c-ea6a034217c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.192998] env[61907]: DEBUG oslo_concurrency.lockutils [req-c69df3e4-3a0c-48f8-84a6-9b64a50acd14 req-fd0e361e-1bdf-426c-bf65-d31607ef1015 service nova] Releasing lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.193607] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquired lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.193717] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.466583] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.789s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.468951] env[61907]: ERROR nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Traceback (most recent call last): [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.driver.spawn(context, instance, image_meta, [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] vm_ref = self.build_virtual_machine(instance, [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.468951] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] for vif in network_info: [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self._sync_wrapper(fn, *args, **kwargs) [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.wait() [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self[:] = self._gt.wait() [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self._exit_event.wait() [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] result = hub.switch() [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.469662] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return self.greenlet.switch() [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] result = function(*args, **kwargs) [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] return func(*args, **kwargs) [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise e [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] nwinfo = self.network_api.allocate_for_instance( [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] created_port_ids = self._update_ports_for_instance( [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] with excutils.save_and_reraise_exception(): [ 565.469976] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] self.force_reraise() [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise self.value [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] updated_port = self._update_port( [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] _ensure_no_port_binding_failure(port) [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] raise exception.PortBindingFailed(port_id=port['id']) [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] nova.exception.PortBindingFailed: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. [ 565.470406] env[61907]: ERROR nova.compute.manager [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] [ 565.470817] env[61907]: DEBUG nova.compute.utils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.470817] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.444s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.472860] env[61907]: INFO nova.compute.claims [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 565.474140] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Build of instance fb4d09e4-6d60-41ed-862c-2bd8eca1e303 was re-scheduled: Binding failed for port ae7efd37-80d8-43d3-a2d3-d60ccdd77a74, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 565.474535] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 565.474781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquiring lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.474932] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Acquired lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.475104] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 565.536315] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 565.760255] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.760574] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.775546] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.045501] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.067567] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.130335] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.264545] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 566.271293] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "0ec61517-a24f-4229-81e5-f889d1871ee7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.271563] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "0ec61517-a24f-4229-81e5-f889d1871ee7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.327345] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.640887] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Releasing lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.640887] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 566.641616] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 566.642455] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c12d2a9-96f3-4182-990f-9aebb780bdc8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.659304] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173403b8-dd50-495e-af7a-af6f9874b3ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.689864] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c45186da-90cb-4551-a7be-4deeae587967 could not be found. [ 566.690143] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 566.690338] env[61907]: INFO nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Took 0.05 seconds to destroy the instance on the hypervisor. [ 566.690930] env[61907]: DEBUG oslo.service.loopingcall [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 566.693451] env[61907]: DEBUG nova.compute.manager [-] [instance: c45186da-90cb-4551-a7be-4deeae587967] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.695013] env[61907]: DEBUG nova.network.neutron [-] [instance: c45186da-90cb-4551-a7be-4deeae587967] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.766782] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c890fbdc-196d-47e4-bdb7-2ac7b36fd086 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.779237] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf8e5473-5457-4244-911f-e24265ae1012 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.819846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.821951] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5bcbcf-6a33-48f3-9cc4-218cdfeb9da6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.828857] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a0c7fc-d600-437a-87b3-3a784929a949 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.833505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Releasing lock "refresh_cache-fb4d09e4-6d60-41ed-862c-2bd8eca1e303" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.833505] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 566.833615] env[61907]: DEBUG nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 566.833852] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 566.846092] env[61907]: DEBUG nova.compute.provider_tree [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.941593] env[61907]: DEBUG nova.network.neutron [-] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 566.981929] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 567.274083] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.274305] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.354324] env[61907]: DEBUG nova.scheduler.client.report [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 567.444814] env[61907]: DEBUG nova.network.neutron [-] [instance: c45186da-90cb-4551-a7be-4deeae587967] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.485804] env[61907]: DEBUG nova.network.neutron [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.596999] env[61907]: DEBUG nova.compute.manager [req-eb2a61f0-ecb5-4a84-80cf-7b8a11a9c0d0 req-23aa9fce-3ed8-4a19-a38b-62eae4eee500 service nova] [instance: c45186da-90cb-4551-a7be-4deeae587967] Received event network-vif-deleted-e37bdea3-0728-41ac-86f4-95ff72e30057 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 567.857872] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.858468] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 567.861262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.459s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.948412] env[61907]: INFO nova.compute.manager [-] [instance: c45186da-90cb-4551-a7be-4deeae587967] Took 1.25 seconds to deallocate network for instance. [ 567.951775] env[61907]: DEBUG nova.compute.claims [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 567.951775] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.991853] env[61907]: INFO nova.compute.manager [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] Took 1.16 seconds to deallocate network for instance. [ 568.365808] env[61907]: DEBUG nova.compute.utils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 568.376242] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 568.376242] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 568.497152] env[61907]: DEBUG nova.policy [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6b2efa7f88504e64b466f502ac6e0540', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ca369b965a942519bb9200e360b0176', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 568.691796] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa132b3c-9a49-4217-ad6e-3141981f999d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.702923] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14bf5bc2-d42f-4203-9def-a71eb7f5b567 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.740130] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ca4910-b3de-4a39-a977-cb54ffebe638 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.748811] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1aaa419-3968-402d-ad79-cfed5f8de166 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.765358] env[61907]: DEBUG nova.compute.provider_tree [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 568.875288] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 569.046237] env[61907]: INFO nova.scheduler.client.report [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Deleted allocations for instance fb4d09e4-6d60-41ed-862c-2bd8eca1e303 [ 569.271834] env[61907]: DEBUG nova.scheduler.client.report [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 569.561243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a636c0f7-9643-4caa-80f0-944d8c2e7a5b tempest-ServerExternalEventsTest-1082564366 tempest-ServerExternalEventsTest-1082564366-project-member] Lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.797s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.562716] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 18.337s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.562716] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: fb4d09e4-6d60-41ed-862c-2bd8eca1e303] During sync_power_state the instance has a pending task (spawning). Skip. [ 569.562881] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "fb4d09e4-6d60-41ed-862c-2bd8eca1e303" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.596778] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Successfully created port: 4bb66af5-840c-4989-a4fd-fcf14ec8f136 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 569.778853] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.918s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.780572] env[61907]: ERROR nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Traceback (most recent call last): [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.driver.spawn(context, instance, image_meta, [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self._vmops.spawn(context, instance, image_meta, injected_files, [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] vm_ref = self.build_virtual_machine(instance, [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] vif_infos = vmwarevif.get_vif_info(self._session, [ 569.780572] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] for vif in network_info: [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self._sync_wrapper(fn, *args, **kwargs) [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.wait() [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self[:] = self._gt.wait() [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self._exit_event.wait() [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] result = hub.switch() [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 569.781019] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return self.greenlet.switch() [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] result = function(*args, **kwargs) [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] return func(*args, **kwargs) [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise e [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] nwinfo = self.network_api.allocate_for_instance( [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] created_port_ids = self._update_ports_for_instance( [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] with excutils.save_and_reraise_exception(): [ 569.781336] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] self.force_reraise() [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise self.value [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] updated_port = self._update_port( [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] _ensure_no_port_binding_failure(port) [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] raise exception.PortBindingFailed(port_id=port['id']) [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] nova.exception.PortBindingFailed: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. [ 569.781643] env[61907]: ERROR nova.compute.manager [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] [ 569.782177] env[61907]: DEBUG nova.compute.utils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 569.787163] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Build of instance a3768960-4967-4975-9760-bc6ceb0b3e64 was re-scheduled: Binding failed for port ed13b448-e776-41b5-a2d8-77507d4d5ddd, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 569.787163] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 569.787163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.787163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.787337] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 569.789685] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.048s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.793932] env[61907]: INFO nova.compute.claims [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.893143] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 569.924213] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 569.924213] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 569.924396] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 569.926764] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 569.926764] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 569.926764] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 569.926764] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 569.926764] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 569.927093] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 569.927093] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 569.927093] env[61907]: DEBUG nova.virt.hardware [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 569.927904] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-043053b2-4aa9-411e-b38e-7955b27608aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.939243] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5d028e-2fd8-4bf6-a2c8-726f07f74163 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.064544] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 570.322315] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.568401] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.600523] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.728162] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "5556aefa-3c0b-415c-94f3-603c7fb26ec1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.728330] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "5556aefa-3c0b-415c-94f3-603c7fb26ec1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.046959] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4637fb-eaf9-4baf-aaa7-137db5297ab9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.058555] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b158c572-722f-46d1-be11-30236992e6e5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.100181] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-a3768960-4967-4975-9760-bc6ceb0b3e64" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.100425] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 571.104016] env[61907]: DEBUG nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 571.104016] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 571.104346] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdcb97f-f149-42e0-b402-ea4bca1e9850 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.115519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a9c36ac-99c7-42fe-8dae-5c6f38d3397b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.132189] env[61907]: DEBUG nova.compute.provider_tree [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.178023] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 571.551114] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquiring lock "bed89649-b8d3-414a-8729-a504a7e02b94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.551265] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "bed89649-b8d3-414a-8729-a504a7e02b94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.633382] env[61907]: ERROR nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 571.633382] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.633382] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.633382] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.633382] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.633382] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.633382] env[61907]: ERROR nova.compute.manager raise self.value [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.633382] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 571.633382] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.633382] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 571.633832] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.633832] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 571.633832] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 571.633832] env[61907]: ERROR nova.compute.manager [ 571.633832] env[61907]: Traceback (most recent call last): [ 571.633832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 571.633832] env[61907]: listener.cb(fileno) [ 571.633832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.633832] env[61907]: result = function(*args, **kwargs) [ 571.633832] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.633832] env[61907]: return func(*args, **kwargs) [ 571.633832] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 571.633832] env[61907]: raise e [ 571.633832] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.633832] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 571.633832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.633832] env[61907]: created_port_ids = self._update_ports_for_instance( [ 571.633832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.633832] env[61907]: with excutils.save_and_reraise_exception(): [ 571.633832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.633832] env[61907]: self.force_reraise() [ 571.633832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.633832] env[61907]: raise self.value [ 571.633832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.633832] env[61907]: updated_port = self._update_port( [ 571.633832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.633832] env[61907]: _ensure_no_port_binding_failure(port) [ 571.633832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.633832] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 571.634527] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 571.634527] env[61907]: Removing descriptor: 15 [ 571.634527] env[61907]: ERROR nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Traceback (most recent call last): [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] yield resources [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.driver.spawn(context, instance, image_meta, [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 571.634527] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] vm_ref = self.build_virtual_machine(instance, [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] for vif in network_info: [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self._sync_wrapper(fn, *args, **kwargs) [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.wait() [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self[:] = self._gt.wait() [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self._exit_event.wait() [ 571.634822] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] result = hub.switch() [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self.greenlet.switch() [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] result = function(*args, **kwargs) [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return func(*args, **kwargs) [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise e [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] nwinfo = self.network_api.allocate_for_instance( [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 571.635130] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] created_port_ids = self._update_ports_for_instance( [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] with excutils.save_and_reraise_exception(): [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.force_reraise() [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise self.value [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] updated_port = self._update_port( [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] _ensure_no_port_binding_failure(port) [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 571.636185] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise exception.PortBindingFailed(port_id=port['id']) [ 571.636685] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 571.636685] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] [ 571.636685] env[61907]: INFO nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Terminating instance [ 571.636844] env[61907]: DEBUG nova.scheduler.client.report [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 571.680550] env[61907]: DEBUG nova.network.neutron [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.143420] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.144275] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 572.149710] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquiring lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.149710] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquired lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.149880] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.150816] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.559s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.184647] env[61907]: INFO nova.compute.manager [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] Took 1.08 seconds to deallocate network for instance. [ 572.653483] env[61907]: DEBUG nova.compute.utils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.657662] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 572.657662] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 572.699403] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 572.779018] env[61907]: DEBUG nova.compute.manager [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Received event network-changed-4bb66af5-840c-4989-a4fd-fcf14ec8f136 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 572.779018] env[61907]: DEBUG nova.compute.manager [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Refreshing instance network info cache due to event network-changed-4bb66af5-840c-4989-a4fd-fcf14ec8f136. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 572.779018] env[61907]: DEBUG oslo_concurrency.lockutils [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] Acquiring lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.783286] env[61907]: DEBUG nova.policy [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2d3c8ea4ad348a48f42bb871a600a5b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fe8cb633b95440b88aa65bb79634251f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.949916] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "c6fd5d45-1abd-4a12-8fcf-86930161bafb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.950176] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "c6fd5d45-1abd-4a12-8fcf-86930161bafb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.950931] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2c2fdf-e391-443b-8c43-d47ccb5ce84c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.959527] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e246c42d-9c3e-4373-9dd9-92a92d5cf750 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.996867] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b28fc924-0c84-4808-918c-9f9bfcb9f1ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.007032] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76c54d7-a4e6-46cd-8b47-efe5370db445 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.019458] env[61907]: DEBUG nova.compute.provider_tree [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.100978] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.163199] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 573.276459] env[61907]: INFO nova.scheduler.client.report [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance a3768960-4967-4975-9760-bc6ceb0b3e64 [ 573.524556] env[61907]: DEBUG nova.scheduler.client.report [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 573.605040] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Releasing lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 573.605040] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 573.605040] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.605040] env[61907]: DEBUG oslo_concurrency.lockutils [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] Acquired lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 573.605451] env[61907]: DEBUG nova.network.neutron [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Refreshing network info cache for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 573.607455] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b34aa214-9705-47c7-903d-490c51f4f184 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.624412] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8361d4-359c-44c2-9048-8c8115fa49ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.660537] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc04ac15-a97a-4248-b578-8f9046ea7e8e could not be found. [ 573.661064] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 573.661363] env[61907]: INFO nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Took 0.06 seconds to destroy the instance on the hypervisor. [ 573.661721] env[61907]: DEBUG oslo.service.loopingcall [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 573.662054] env[61907]: DEBUG nova.compute.manager [-] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 573.662384] env[61907]: DEBUG nova.network.neutron [-] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 573.698888] env[61907]: DEBUG nova.network.neutron [-] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.784274] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Successfully created port: 52d682ea-ac66-4d00-83f7-171ac8a91d72 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.786583] env[61907]: DEBUG oslo_concurrency.lockutils [None req-01045ccf-ae21-4550-b3af-9e462f5044b0 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "a3768960-4967-4975-9760-bc6ceb0b3e64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.170s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.787680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "a3768960-4967-4975-9760-bc6ceb0b3e64" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 22.562s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.787937] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a3768960-4967-4975-9760-bc6ceb0b3e64] During sync_power_state the instance has a pending task (spawning). Skip. [ 573.788017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "a3768960-4967-4975-9760-bc6ceb0b3e64" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.035111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.882s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.035111] env[61907]: ERROR nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Traceback (most recent call last): [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.driver.spawn(context, instance, image_meta, [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.035111] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] vm_ref = self.build_virtual_machine(instance, [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] for vif in network_info: [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self._sync_wrapper(fn, *args, **kwargs) [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.wait() [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self[:] = self._gt.wait() [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self._exit_event.wait() [ 574.035623] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] result = hub.switch() [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return self.greenlet.switch() [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] result = function(*args, **kwargs) [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] return func(*args, **kwargs) [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise e [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] nwinfo = self.network_api.allocate_for_instance( [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 574.035984] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] created_port_ids = self._update_ports_for_instance( [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] with excutils.save_and_reraise_exception(): [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] self.force_reraise() [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise self.value [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] updated_port = self._update_port( [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] _ensure_no_port_binding_failure(port) [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.036334] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] raise exception.PortBindingFailed(port_id=port['id']) [ 574.036684] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] nova.exception.PortBindingFailed: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. [ 574.036684] env[61907]: ERROR nova.compute.manager [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] [ 574.036684] env[61907]: DEBUG nova.compute.utils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 574.038972] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Build of instance eb3a4f2a-8667-4d05-8606-a2dcb93c94a2 was re-scheduled: Binding failed for port 4aa89582-e057-45e2-ab6a-da8aca9f219c, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 574.038972] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 574.038972] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquiring lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.038972] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Acquired lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.039287] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 574.042074] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.519s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.149778] env[61907]: DEBUG nova.network.neutron [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.182719] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 574.205371] env[61907]: DEBUG nova.network.neutron [-] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.220198] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.220198] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.220198] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.220380] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.220380] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.220380] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.220380] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.221119] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.221417] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.221607] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.221790] env[61907]: DEBUG nova.virt.hardware [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.223519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f6620a-f188-4de6-84fe-4e7a5a411cc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.235575] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade2f361-46fe-4e83-80b9-4f3d7789fe7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.290013] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 574.332885] env[61907]: DEBUG nova.network.neutron [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.333130] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquiring lock "2fcb4164-c0c5-49e5-87cb-a51df711b5a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.333387] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "2fcb4164-c0c5-49e5-87cb-a51df711b5a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.603827] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.710144] env[61907]: INFO nova.compute.manager [-] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Took 1.05 seconds to deallocate network for instance. [ 574.719038] env[61907]: DEBUG nova.compute.claims [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 574.719038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.823191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 574.835173] env[61907]: DEBUG oslo_concurrency.lockutils [req-89a53c76-5634-4bbf-a47c-bb018b0cc384 req-1c3b1603-3eb7-4f8b-a10a-74a169876979 service nova] Releasing lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.880110] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc943f4-be76-4eae-9dda-3f8bb5a9b7fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.889094] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b990e066-2ea1-41cd-b802-f8de963b9478 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.924489] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d49267e5-b2c3-4fe9-a387-c09520636e5e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.933822] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64e1e3f-eaaa-4b17-b869-17ba7243fed3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.952184] env[61907]: DEBUG nova.compute.provider_tree [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.972934] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 575.454569] env[61907]: DEBUG nova.scheduler.client.report [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 575.475505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Releasing lock "refresh_cache-eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 575.475602] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 575.476707] env[61907]: DEBUG nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 575.476707] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 575.529532] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 575.963642] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.923s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.964591] env[61907]: ERROR nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] Traceback (most recent call last): [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.driver.spawn(context, instance, image_meta, [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] vm_ref = self.build_virtual_machine(instance, [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.964591] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] for vif in network_info: [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self._sync_wrapper(fn, *args, **kwargs) [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.wait() [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self[:] = self._gt.wait() [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self._exit_event.wait() [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] result = hub.switch() [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.965137] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return self.greenlet.switch() [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] result = function(*args, **kwargs) [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] return func(*args, **kwargs) [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise e [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] nwinfo = self.network_api.allocate_for_instance( [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] created_port_ids = self._update_ports_for_instance( [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] with excutils.save_and_reraise_exception(): [ 575.965603] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] self.force_reraise() [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise self.value [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] updated_port = self._update_port( [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] _ensure_no_port_binding_failure(port) [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] raise exception.PortBindingFailed(port_id=port['id']) [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] nova.exception.PortBindingFailed: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. [ 575.966068] env[61907]: ERROR nova.compute.manager [instance: a58bebcc-73e2-4752-befb-91083baacabf] [ 575.966476] env[61907]: DEBUG nova.compute.utils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.966606] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.443s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.968198] env[61907]: INFO nova.compute.claims [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.973362] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Build of instance a58bebcc-73e2-4752-befb-91083baacabf was re-scheduled: Binding failed for port a2cc7c85-6d99-4eb1-85aa-023d902efaef, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 575.973824] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 575.974191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquiring lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.974605] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Acquired lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.974847] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.034149] env[61907]: DEBUG nova.network.neutron [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.238503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquiring lock "60d87612-650c-48dd-8afb-e0dc181c9c70" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.238818] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "60d87612-650c-48dd-8afb-e0dc181c9c70" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.475999] env[61907]: ERROR nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 576.475999] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.475999] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.475999] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.475999] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.475999] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.475999] env[61907]: ERROR nova.compute.manager raise self.value [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.475999] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.475999] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.475999] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.476769] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.476769] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.476769] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 576.476769] env[61907]: ERROR nova.compute.manager [ 576.476769] env[61907]: Traceback (most recent call last): [ 576.476769] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.476769] env[61907]: listener.cb(fileno) [ 576.476769] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.476769] env[61907]: result = function(*args, **kwargs) [ 576.476769] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.476769] env[61907]: return func(*args, **kwargs) [ 576.476769] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.476769] env[61907]: raise e [ 576.476769] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.476769] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 576.476769] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.476769] env[61907]: created_port_ids = self._update_ports_for_instance( [ 576.476769] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.476769] env[61907]: with excutils.save_and_reraise_exception(): [ 576.476769] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.476769] env[61907]: self.force_reraise() [ 576.476769] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.476769] env[61907]: raise self.value [ 576.476769] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.476769] env[61907]: updated_port = self._update_port( [ 576.476769] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.476769] env[61907]: _ensure_no_port_binding_failure(port) [ 576.476769] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.476769] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.477617] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 576.477617] env[61907]: Removing descriptor: 15 [ 576.477617] env[61907]: ERROR nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Traceback (most recent call last): [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] yield resources [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.driver.spawn(context, instance, image_meta, [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.477617] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] vm_ref = self.build_virtual_machine(instance, [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] for vif in network_info: [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self._sync_wrapper(fn, *args, **kwargs) [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.wait() [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self[:] = self._gt.wait() [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self._exit_event.wait() [ 576.477970] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] result = hub.switch() [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self.greenlet.switch() [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] result = function(*args, **kwargs) [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return func(*args, **kwargs) [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise e [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] nwinfo = self.network_api.allocate_for_instance( [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.478383] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] created_port_ids = self._update_ports_for_instance( [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] with excutils.save_and_reraise_exception(): [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.force_reraise() [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise self.value [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] updated_port = self._update_port( [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] _ensure_no_port_binding_failure(port) [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.478764] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise exception.PortBindingFailed(port_id=port['id']) [ 576.479134] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 576.479134] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] [ 576.479134] env[61907]: INFO nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Terminating instance [ 576.512525] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.522279] env[61907]: DEBUG nova.compute.manager [req-b83871b3-bb6d-4afb-83b2-fca704cae1a2 req-e32d218b-acd5-46b0-9758-7dfb059a236e service nova] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Received event network-vif-deleted-4bb66af5-840c-4989-a4fd-fcf14ec8f136 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 576.542491] env[61907]: INFO nova.compute.manager [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] Took 1.07 seconds to deallocate network for instance. [ 576.685644] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.989564] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquiring lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.989803] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquired lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.989981] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.191668] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Releasing lock "refresh_cache-a58bebcc-73e2-4752-befb-91083baacabf" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.191924] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 577.192304] env[61907]: DEBUG nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 577.192370] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.226415] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.306946] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ac9ea8-f955-4906-b975-c1d61e1ae89b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.315248] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f46f00-392e-4f9a-ad6e-ae08c44cf232 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.356733] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba9ad76-b9c3-4784-a75b-98c231090087 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.366451] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ab286cf-e39f-4192-b887-b548e54c406f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.385381] env[61907]: DEBUG nova.compute.provider_tree [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.521596] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.545501] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "92fa087f-a147-438b-9c30-c93baa5828dc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.545501] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "92fa087f-a147-438b-9c30-c93baa5828dc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.592666] env[61907]: INFO nova.scheduler.client.report [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Deleted allocations for instance eb3a4f2a-8667-4d05-8606-a2dcb93c94a2 [ 577.733198] env[61907]: DEBUG nova.network.neutron [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.745970] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.893525] env[61907]: DEBUG nova.scheduler.client.report [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 578.105024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4404c4a7-0377-4919-9ef4-85174b7b2ce6 tempest-FloatingIPsAssociationTestJSON-1794669245 tempest-FloatingIPsAssociationTestJSON-1794669245-project-member] Lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.649s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.105024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 26.878s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.105024] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-905f0a93-4b0f-4e50-8f5a-61148fba4fe4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.116110] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07fc2d28-0449-4beb-bf24-59b36f574c8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.236744] env[61907]: INFO nova.compute.manager [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] [instance: a58bebcc-73e2-4752-befb-91083baacabf] Took 1.04 seconds to deallocate network for instance. [ 578.247665] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Releasing lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.248131] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 578.248341] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 578.248626] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff2cfb98-c427-42f6-a1f8-cac042e6056a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.264444] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c15dbc9a-f08a-4e97-8ced-928069931eb7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.296136] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4db46e0e-1aaa-4168-8365-6dab93eee759 could not be found. [ 578.296296] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 578.296478] env[61907]: INFO nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Took 0.05 seconds to destroy the instance on the hypervisor. [ 578.296731] env[61907]: DEBUG oslo.service.loopingcall [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 578.297664] env[61907]: DEBUG nova.compute.manager [-] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 578.297744] env[61907]: DEBUG nova.network.neutron [-] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.336087] env[61907]: DEBUG nova.network.neutron [-] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.401345] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.401345] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 578.402017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.335s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.403394] env[61907]: INFO nova.compute.claims [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 578.562820] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquiring lock "420a661e-33aa-43e5-b298-19a20e8ab33b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.563175] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "420a661e-33aa-43e5-b298-19a20e8ab33b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.609544] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 578.647297] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: eb3a4f2a-8667-4d05-8606-a2dcb93c94a2] During the sync_power process the instance has moved from host None to host cpu-1 [ 578.647491] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "eb3a4f2a-8667-4d05-8606-a2dcb93c94a2" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.543s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.843575] env[61907]: DEBUG nova.network.neutron [-] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.907491] env[61907]: DEBUG nova.compute.utils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.915179] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 578.915179] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 579.038131] env[61907]: DEBUG nova.policy [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5d7de380a6154327bdb18ece30ee3e6c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dd6b0bde4bf6419f8fb789c179b2b0fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.134679] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.281126] env[61907]: INFO nova.scheduler.client.report [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Deleted allocations for instance a58bebcc-73e2-4752-befb-91083baacabf [ 579.345040] env[61907]: INFO nova.compute.manager [-] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Took 1.05 seconds to deallocate network for instance. [ 579.349138] env[61907]: DEBUG nova.compute.claims [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 579.349450] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.415169] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 579.552125] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquiring lock "a7e883a1-aab9-4372-8379-144b373062e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.552125] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "a7e883a1-aab9-4372-8379-144b373062e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.559193] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "d95c4916-660e-4ead-ba87-2f69871d5f1e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.559530] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "d95c4916-660e-4ead-ba87-2f69871d5f1e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.789811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-045be7a7-8553-4c25-8653-9c02cc8ab299 tempest-ServerDiagnosticsNegativeTest-788079183 tempest-ServerDiagnosticsNegativeTest-788079183-project-member] Lock "a58bebcc-73e2-4752-befb-91083baacabf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.426s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.812696] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c545b6-c3bf-4bb4-80d1-33a5ae95c5a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.822090] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dccb95e-3c5a-4c47-b690-222fb023835f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.865085] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf25036-1d97-481e-b119-b23f5937405e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.875201] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c61462-7b88-408c-8ac7-a23f2324cf63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.902154] env[61907]: DEBUG nova.compute.provider_tree [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.908630] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.908813] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.067016] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Successfully created port: 7205218f-561b-4da0-9df1-4aa046a770aa {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 580.185807] env[61907]: DEBUG nova.compute.manager [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Received event network-changed-52d682ea-ac66-4d00-83f7-171ac8a91d72 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 580.185807] env[61907]: DEBUG nova.compute.manager [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Refreshing instance network info cache due to event network-changed-52d682ea-ac66-4d00-83f7-171ac8a91d72. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 580.186043] env[61907]: DEBUG oslo_concurrency.lockutils [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] Acquiring lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.186161] env[61907]: DEBUG oslo_concurrency.lockutils [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] Acquired lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.186340] env[61907]: DEBUG nova.network.neutron [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Refreshing network info cache for port 52d682ea-ac66-4d00-83f7-171ac8a91d72 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.295231] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 580.409100] env[61907]: DEBUG nova.scheduler.client.report [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 580.437322] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 580.469917] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:07:25Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1697724550',id=21,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-2111115336',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.470854] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.470854] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.471115] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.471225] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.471770] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.471770] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.471770] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.472050] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.472050] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.472172] env[61907]: DEBUG nova.virt.hardware [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.473700] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f4bc06-86f2-4a13-9292-feb6b180f801 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.485456] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29c44bb-66ab-47e8-a0b9-d13fa19c8630 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.743182] env[61907]: DEBUG nova.network.neutron [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.831052] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.919238] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.919238] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 580.920026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.100s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.927176] env[61907]: INFO nova.compute.claims [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 581.029373] env[61907]: DEBUG nova.network.neutron [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.432930] env[61907]: DEBUG nova.compute.utils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 581.444803] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 581.445067] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 581.532469] env[61907]: DEBUG oslo_concurrency.lockutils [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] Releasing lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.532469] env[61907]: DEBUG nova.compute.manager [req-3710915f-0565-42e2-a07a-e5fa853cb91f req-c0e46ca4-d509-479d-ab4a-3ca34027c16c service nova] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Received event network-vif-deleted-52d682ea-ac66-4d00-83f7-171ac8a91d72 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 581.582576] env[61907]: DEBUG nova.policy [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1ce727e3e304823abca558d525b59c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '756f32a1aae0494aaf016825d22de3db', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 581.942544] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 582.265302] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Successfully created port: b15d2a73-50b5-4506-9d63-68aeef31e1d0 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 582.285959] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c577bb-f1e6-4256-9b2f-21528f67c4cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.297378] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-221ec1d6-c8cf-4f49-b662-2901fe4b6c17 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.337096] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37ae81f-ff50-4f0e-9dc9-d36fd0cecec6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.342972] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5817086d-3d67-487e-9ac5-fe54cc767435 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.364913] env[61907]: DEBUG nova.compute.provider_tree [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.872877] env[61907]: DEBUG nova.scheduler.client.report [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 582.936843] env[61907]: ERROR nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 582.936843] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.936843] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.936843] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.936843] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.936843] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.936843] env[61907]: ERROR nova.compute.manager raise self.value [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.936843] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.936843] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.936843] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.937635] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.937635] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.937635] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 582.937635] env[61907]: ERROR nova.compute.manager [ 582.937635] env[61907]: Traceback (most recent call last): [ 582.937635] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.937635] env[61907]: listener.cb(fileno) [ 582.937635] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.937635] env[61907]: result = function(*args, **kwargs) [ 582.937635] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.937635] env[61907]: return func(*args, **kwargs) [ 582.937635] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.937635] env[61907]: raise e [ 582.937635] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.937635] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 582.937635] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.937635] env[61907]: created_port_ids = self._update_ports_for_instance( [ 582.937635] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.937635] env[61907]: with excutils.save_and_reraise_exception(): [ 582.937635] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.937635] env[61907]: self.force_reraise() [ 582.937635] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.937635] env[61907]: raise self.value [ 582.937635] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.937635] env[61907]: updated_port = self._update_port( [ 582.937635] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.937635] env[61907]: _ensure_no_port_binding_failure(port) [ 582.937635] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.937635] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.938548] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 582.938548] env[61907]: Removing descriptor: 15 [ 582.938548] env[61907]: ERROR nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Traceback (most recent call last): [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] yield resources [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.driver.spawn(context, instance, image_meta, [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.938548] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] vm_ref = self.build_virtual_machine(instance, [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] for vif in network_info: [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self._sync_wrapper(fn, *args, **kwargs) [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.wait() [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self[:] = self._gt.wait() [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self._exit_event.wait() [ 582.939176] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] result = hub.switch() [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self.greenlet.switch() [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] result = function(*args, **kwargs) [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return func(*args, **kwargs) [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise e [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] nwinfo = self.network_api.allocate_for_instance( [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.939591] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] created_port_ids = self._update_ports_for_instance( [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] with excutils.save_and_reraise_exception(): [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.force_reraise() [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise self.value [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] updated_port = self._update_port( [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] _ensure_no_port_binding_failure(port) [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.940095] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise exception.PortBindingFailed(port_id=port['id']) [ 582.940958] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 582.940958] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] [ 582.940958] env[61907]: INFO nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Terminating instance [ 582.955903] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 582.988710] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.989450] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.989450] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.989450] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.989450] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.989873] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.990206] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.990416] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.990603] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.990771] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.990952] env[61907]: DEBUG nova.virt.hardware [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.992706] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17cb8ee-4aed-4a72-9d81-680199fcc856 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.004309] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf2552a4-d31f-4ea7-b527-813136b883e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.379173] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.380157] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 583.383471] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.432s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.395786] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquiring lock "0da17f71-46ee-42f3-99f7-62f4db6e1d9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.396081] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "0da17f71-46ee-42f3-99f7-62f4db6e1d9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.447263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquiring lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.447448] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquired lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.447624] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 583.890309] env[61907]: DEBUG nova.compute.utils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.897102] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 583.897102] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.982869] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.066945] env[61907]: DEBUG nova.policy [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd9e3b2602d342d0ba69252be126ffa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7bd135f842743baa63fd1ac28416ad0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 584.198545] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.239476] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a80cfd-47dc-4a73-9121-dc3fdde0e780 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.247425] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1554a2-9490-4f31-bbdd-74dd42f16da2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.282263] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749bbb88-7c6c-43e3-83cb-0ec35bca8899 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.291309] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764d89b8-3d92-4358-9c57-d809ea9501c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.307146] env[61907]: DEBUG nova.compute.provider_tree [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.397764] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 584.439111] env[61907]: DEBUG nova.compute.manager [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Received event network-changed-7205218f-561b-4da0-9df1-4aa046a770aa {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 584.440934] env[61907]: DEBUG nova.compute.manager [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Refreshing instance network info cache due to event network-changed-7205218f-561b-4da0-9df1-4aa046a770aa. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 584.441212] env[61907]: DEBUG oslo_concurrency.lockutils [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] Acquiring lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.674158] env[61907]: ERROR nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 584.674158] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.674158] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.674158] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.674158] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.674158] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.674158] env[61907]: ERROR nova.compute.manager raise self.value [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.674158] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.674158] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.674158] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.674618] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.674618] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.674618] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 584.674618] env[61907]: ERROR nova.compute.manager [ 584.674618] env[61907]: Traceback (most recent call last): [ 584.674618] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.674618] env[61907]: listener.cb(fileno) [ 584.674618] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.674618] env[61907]: result = function(*args, **kwargs) [ 584.674618] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.674618] env[61907]: return func(*args, **kwargs) [ 584.674618] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.674618] env[61907]: raise e [ 584.674618] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.674618] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 584.674618] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.674618] env[61907]: created_port_ids = self._update_ports_for_instance( [ 584.674618] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.674618] env[61907]: with excutils.save_and_reraise_exception(): [ 584.674618] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.674618] env[61907]: self.force_reraise() [ 584.674618] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.674618] env[61907]: raise self.value [ 584.674618] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.674618] env[61907]: updated_port = self._update_port( [ 584.674618] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.674618] env[61907]: _ensure_no_port_binding_failure(port) [ 584.674618] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.674618] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.675411] env[61907]: nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 584.675411] env[61907]: Removing descriptor: 17 [ 584.675411] env[61907]: ERROR nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Traceback (most recent call last): [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] yield resources [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.driver.spawn(context, instance, image_meta, [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.675411] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] vm_ref = self.build_virtual_machine(instance, [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] for vif in network_info: [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self._sync_wrapper(fn, *args, **kwargs) [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.wait() [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self[:] = self._gt.wait() [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self._exit_event.wait() [ 584.675738] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] result = hub.switch() [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self.greenlet.switch() [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] result = function(*args, **kwargs) [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return func(*args, **kwargs) [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise e [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] nwinfo = self.network_api.allocate_for_instance( [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.676158] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] created_port_ids = self._update_ports_for_instance( [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] with excutils.save_and_reraise_exception(): [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.force_reraise() [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise self.value [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] updated_port = self._update_port( [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] _ensure_no_port_binding_failure(port) [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.676560] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise exception.PortBindingFailed(port_id=port['id']) [ 584.676888] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 584.676888] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] [ 584.676888] env[61907]: INFO nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Terminating instance [ 584.682499] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "ab7d4c93-fbc9-4446-ac57-f479ab408180" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.682722] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "ab7d4c93-fbc9-4446-ac57-f479ab408180" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.706803] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Releasing lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.707488] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 584.707759] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 584.708164] env[61907]: DEBUG oslo_concurrency.lockutils [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] Acquired lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.708312] env[61907]: DEBUG nova.network.neutron [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Refreshing network info cache for port 7205218f-561b-4da0-9df1-4aa046a770aa {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.709673] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1637577d-9036-4c5b-b783-5907642440e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.720454] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d1a2dbd-b015-4c46-b9a6-10194d1947c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.747450] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71013ca4-f1d0-4110-8fad-91e619ff354b could not be found. [ 584.747450] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.747450] env[61907]: INFO nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 584.747952] env[61907]: DEBUG oslo.service.loopingcall [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.747952] env[61907]: DEBUG nova.compute.manager [-] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 584.748176] env[61907]: DEBUG nova.network.neutron [-] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.810476] env[61907]: DEBUG nova.scheduler.client.report [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 584.914792] env[61907]: DEBUG nova.network.neutron [-] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.166734] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Successfully created port: df82cc2c-0312-46b4-813a-318ad3ff342d {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 585.186283] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquiring lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.186824] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquired lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.186824] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.245947] env[61907]: DEBUG nova.network.neutron [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.315438] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.932s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.316267] env[61907]: ERROR nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] Traceback (most recent call last): [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.driver.spawn(context, instance, image_meta, [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] vm_ref = self.build_virtual_machine(instance, [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.316267] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] for vif in network_info: [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self._sync_wrapper(fn, *args, **kwargs) [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.wait() [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self[:] = self._gt.wait() [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self._exit_event.wait() [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] result = hub.switch() [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.316627] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return self.greenlet.switch() [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] result = function(*args, **kwargs) [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] return func(*args, **kwargs) [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise e [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] nwinfo = self.network_api.allocate_for_instance( [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] created_port_ids = self._update_ports_for_instance( [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] with excutils.save_and_reraise_exception(): [ 585.316974] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] self.force_reraise() [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise self.value [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] updated_port = self._update_port( [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] _ensure_no_port_binding_failure(port) [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] raise exception.PortBindingFailed(port_id=port['id']) [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] nova.exception.PortBindingFailed: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. [ 585.317336] env[61907]: ERROR nova.compute.manager [instance: c45186da-90cb-4551-a7be-4deeae587967] [ 585.320568] env[61907]: DEBUG nova.compute.utils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 585.320568] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.718s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.320568] env[61907]: INFO nova.compute.claims [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.324036] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Build of instance c45186da-90cb-4551-a7be-4deeae587967 was re-scheduled: Binding failed for port e37bdea3-0728-41ac-86f4-95ff72e30057, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 585.324036] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 585.324170] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquiring lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.324237] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Acquired lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.324372] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 585.407570] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 585.418612] env[61907]: DEBUG nova.network.neutron [-] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.457142] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.457421] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.457583] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.457769] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.457916] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.458106] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.458353] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.458523] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.458695] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.459075] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.459291] env[61907]: DEBUG nova.virt.hardware [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.460504] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c362d9-87ab-4490-ac73-8a187f964dd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.471440] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5840b859-4e1f-4c13-b97b-d0865e5f74b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.538426] env[61907]: DEBUG nova.network.neutron [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.726268] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.834160] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.921036] env[61907]: INFO nova.compute.manager [-] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Took 1.17 seconds to deallocate network for instance. [ 585.923949] env[61907]: DEBUG nova.compute.claims [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 585.924267] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.968194] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.041407] env[61907]: DEBUG oslo_concurrency.lockutils [req-e6a3b704-1bd8-4c57-9334-7f41a2a233ce req-e928c00a-2f15-42ba-ab5d-8123645896b9 service nova] Releasing lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.139448] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.344106] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Releasing lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.344106] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 586.344106] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 586.344106] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-645bf503-fa96-4f85-a096-ed042bfabe0e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.356158] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea01b04-7fdd-43e0-991f-b66eb59f0975 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.389498] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f66cffbb-1af4-4675-ae4c-ea6a034217c7 could not be found. [ 586.389498] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 586.389498] env[61907]: INFO nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 586.389498] env[61907]: DEBUG oslo.service.loopingcall [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.389498] env[61907]: DEBUG nova.compute.manager [-] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.389498] env[61907]: DEBUG nova.network.neutron [-] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.424693] env[61907]: DEBUG nova.network.neutron [-] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.641732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Releasing lock "refresh_cache-c45186da-90cb-4551-a7be-4deeae587967" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.642013] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 586.642238] env[61907]: DEBUG nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 586.642353] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 586.670938] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.674821] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b65069f-d7c6-4d76-8df5-86e61847af42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.685427] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb78b20a-181f-4014-a56e-0fc306de5901 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.722027] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40835268-8299-460b-8724-84756d181477 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.728654] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d08749-b565-4a1a-8845-eeedfeaca471 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.744647] env[61907]: DEBUG nova.compute.provider_tree [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.926513] env[61907]: DEBUG nova.network.neutron [-] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.093309] env[61907]: ERROR nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 587.093309] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.093309] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.093309] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.093309] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.093309] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.093309] env[61907]: ERROR nova.compute.manager raise self.value [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.093309] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.093309] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.093309] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.093784] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.093784] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.093784] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 587.093784] env[61907]: ERROR nova.compute.manager [ 587.093784] env[61907]: Traceback (most recent call last): [ 587.093784] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.093784] env[61907]: listener.cb(fileno) [ 587.093784] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.093784] env[61907]: result = function(*args, **kwargs) [ 587.093784] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.093784] env[61907]: return func(*args, **kwargs) [ 587.093784] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 587.093784] env[61907]: raise e [ 587.093784] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.093784] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 587.093784] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.093784] env[61907]: created_port_ids = self._update_ports_for_instance( [ 587.093784] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.093784] env[61907]: with excutils.save_and_reraise_exception(): [ 587.093784] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.093784] env[61907]: self.force_reraise() [ 587.093784] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.093784] env[61907]: raise self.value [ 587.093784] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.093784] env[61907]: updated_port = self._update_port( [ 587.093784] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.093784] env[61907]: _ensure_no_port_binding_failure(port) [ 587.093784] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.093784] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.094767] env[61907]: nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 587.094767] env[61907]: Removing descriptor: 15 [ 587.094767] env[61907]: ERROR nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Traceback (most recent call last): [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] yield resources [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.driver.spawn(context, instance, image_meta, [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.094767] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] vm_ref = self.build_virtual_machine(instance, [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] for vif in network_info: [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self._sync_wrapper(fn, *args, **kwargs) [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.wait() [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self[:] = self._gt.wait() [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self._exit_event.wait() [ 587.095148] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] result = hub.switch() [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self.greenlet.switch() [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] result = function(*args, **kwargs) [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return func(*args, **kwargs) [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise e [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] nwinfo = self.network_api.allocate_for_instance( [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.095549] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] created_port_ids = self._update_ports_for_instance( [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] with excutils.save_and_reraise_exception(): [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.force_reraise() [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise self.value [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] updated_port = self._update_port( [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] _ensure_no_port_binding_failure(port) [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.095909] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise exception.PortBindingFailed(port_id=port['id']) [ 587.096256] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 587.096256] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] [ 587.096256] env[61907]: INFO nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Terminating instance [ 587.180105] env[61907]: DEBUG nova.network.neutron [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.248348] env[61907]: DEBUG nova.scheduler.client.report [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 587.342487] env[61907]: DEBUG nova.compute.manager [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Received event network-vif-deleted-7205218f-561b-4da0-9df1-4aa046a770aa {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 587.346614] env[61907]: DEBUG nova.compute.manager [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Received event network-changed-b15d2a73-50b5-4506-9d63-68aeef31e1d0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 587.346976] env[61907]: DEBUG nova.compute.manager [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Refreshing instance network info cache due to event network-changed-b15d2a73-50b5-4506-9d63-68aeef31e1d0. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 587.347192] env[61907]: DEBUG oslo_concurrency.lockutils [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] Acquiring lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.347379] env[61907]: DEBUG oslo_concurrency.lockutils [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] Acquired lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.347516] env[61907]: DEBUG nova.network.neutron [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Refreshing network info cache for port b15d2a73-50b5-4506-9d63-68aeef31e1d0 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 587.431130] env[61907]: INFO nova.compute.manager [-] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Took 1.04 seconds to deallocate network for instance. [ 587.432097] env[61907]: DEBUG nova.compute.claims [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.432278] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.601854] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.603126] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquired lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.603343] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.683550] env[61907]: INFO nova.compute.manager [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] [instance: c45186da-90cb-4551-a7be-4deeae587967] Took 1.04 seconds to deallocate network for instance. [ 587.755680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.755680] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 587.758082] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.039s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.915149] env[61907]: DEBUG nova.network.neutron [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.142724] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquiring lock "78f3d457-d76b-4a13-9f12-264e53b768b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.143382] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "78f3d457-d76b-4a13-9f12-264e53b768b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.147710] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.163465] env[61907]: DEBUG nova.network.neutron [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.265275] env[61907]: DEBUG nova.compute.utils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.270842] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 588.271073] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.353912] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.600596] env[61907]: DEBUG nova.policy [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '10d6100ad10f4e6e93914e3cfa923aa8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '75a29d8205814a0b8ae2d5a45705312f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.630399] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280a77ef-b6da-45df-9a10-9eaf1d56f588 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.640416] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e7a143-ac06-4fe6-aef4-5644d91be451 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.687871] env[61907]: DEBUG oslo_concurrency.lockutils [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] Releasing lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.687981] env[61907]: DEBUG nova.compute.manager [req-6ee29f04-a1a9-418a-bbfb-39d3b550e18c req-6b1522a4-91fd-4bc5-ac7f-d9110a8060df service nova] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Received event network-vif-deleted-b15d2a73-50b5-4506-9d63-68aeef31e1d0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 588.689924] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34fe7e85-7808-4209-975e-3eb0c654bca2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.704874] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d31977d-d94b-4b26-a2d6-08abd897d2ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.726295] env[61907]: DEBUG nova.compute.provider_tree [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.731589] env[61907]: INFO nova.scheduler.client.report [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Deleted allocations for instance c45186da-90cb-4551-a7be-4deeae587967 [ 588.772226] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 588.856569] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Releasing lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.858368] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 588.858674] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 588.858992] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-537fa35b-1408-4eda-8cb5-1c64f6731bff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.871942] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc3af5b-61fc-410a-b030-855f7b129f00 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.898928] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b3937fb5-74f4-4832-b9cf-9c51ee8ceb50 could not be found. [ 588.899207] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 588.899412] env[61907]: INFO nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Took 0.04 seconds to destroy the instance on the hypervisor. [ 588.899631] env[61907]: DEBUG oslo.service.loopingcall [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 588.899830] env[61907]: DEBUG nova.compute.manager [-] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 588.899931] env[61907]: DEBUG nova.network.neutron [-] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.017022] env[61907]: DEBUG nova.network.neutron [-] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.231351] env[61907]: DEBUG nova.scheduler.client.report [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 589.240530] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebc6a13b-c549-4846-b6ce-0a041e74dc74 tempest-TenantUsagesTestJSON-153348568 tempest-TenantUsagesTestJSON-153348568-project-member] Lock "c45186da-90cb-4551-a7be-4deeae587967" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.736s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.520753] env[61907]: DEBUG nova.network.neutron [-] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.682103] env[61907]: DEBUG nova.compute.manager [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Received event network-changed-df82cc2c-0312-46b4-813a-318ad3ff342d {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 589.682707] env[61907]: DEBUG nova.compute.manager [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Refreshing instance network info cache due to event network-changed-df82cc2c-0312-46b4-813a-318ad3ff342d. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 589.682707] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] Acquiring lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.682837] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] Acquired lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.682937] env[61907]: DEBUG nova.network.neutron [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Refreshing network info cache for port df82cc2c-0312-46b4-813a-318ad3ff342d {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 589.741782] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.984s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.742387] env[61907]: ERROR nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Traceback (most recent call last): [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.driver.spawn(context, instance, image_meta, [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] vm_ref = self.build_virtual_machine(instance, [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.742387] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] for vif in network_info: [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self._sync_wrapper(fn, *args, **kwargs) [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.wait() [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self[:] = self._gt.wait() [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self._exit_event.wait() [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] result = hub.switch() [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.742713] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return self.greenlet.switch() [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] result = function(*args, **kwargs) [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] return func(*args, **kwargs) [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise e [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] nwinfo = self.network_api.allocate_for_instance( [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] created_port_ids = self._update_ports_for_instance( [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] with excutils.save_and_reraise_exception(): [ 589.743121] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] self.force_reraise() [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise self.value [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] updated_port = self._update_port( [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] _ensure_no_port_binding_failure(port) [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] raise exception.PortBindingFailed(port_id=port['id']) [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] nova.exception.PortBindingFailed: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. [ 589.743470] env[61907]: ERROR nova.compute.manager [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] [ 589.743762] env[61907]: DEBUG nova.compute.utils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.744465] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.924s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.745903] env[61907]: INFO nova.compute.claims [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.748540] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 589.755016] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Build of instance dc04ac15-a97a-4248-b578-8f9046ea7e8e was re-scheduled: Binding failed for port 4bb66af5-840c-4989-a4fd-fcf14ec8f136, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 589.755016] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 589.755016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquiring lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.755016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Acquired lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.755255] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.783712] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 589.823589] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.823840] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.823994] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.824245] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.824333] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.824478] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.824822] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.824822] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.825304] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.825601] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.825818] env[61907]: DEBUG nova.virt.hardware [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.826855] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1815652b-aee0-4b17-8fde-9798e3c21285 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.836789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55bff52-eecf-4dc7-a67d-716921809fcf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.025714] env[61907]: INFO nova.compute.manager [-] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Took 1.13 seconds to deallocate network for instance. [ 590.029270] env[61907]: DEBUG nova.compute.claims [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.030126] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.226075] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Successfully created port: e1d0935d-af16-44a7-ad0e-1c8aa97c7074 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.228904] env[61907]: DEBUG nova.network.neutron [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.284976] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.315418] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.413375] env[61907]: DEBUG nova.network.neutron [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.620511] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.916131] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] Releasing lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.916380] env[61907]: DEBUG nova.compute.manager [req-bf9c7629-0878-4a01-9c33-06a70ce22b9b req-46dc2fac-782a-4a08-9209-46107852b36f service nova] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Received event network-vif-deleted-df82cc2c-0312-46b4-813a-318ad3ff342d {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 591.126233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Releasing lock "refresh_cache-dc04ac15-a97a-4248-b578-8f9046ea7e8e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.126631] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 591.126921] env[61907]: DEBUG nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 591.127348] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.142364] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fd3d402-8360-4cd9-aa0f-a78f9c1a5d28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.155536] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f1e862-e15b-4785-9d0b-9925da27c589 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.190672] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5445f30d-6005-4a56-b0ec-cf73aa399e0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.196290] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.202951] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3efcc672-fd61-4fa1-af13-f5d649209a1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.221290] env[61907]: DEBUG nova.compute.provider_tree [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.697809] env[61907]: DEBUG nova.network.neutron [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.727036] env[61907]: DEBUG nova.scheduler.client.report [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 592.202335] env[61907]: INFO nova.compute.manager [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] [instance: dc04ac15-a97a-4248-b578-8f9046ea7e8e] Took 1.07 seconds to deallocate network for instance. [ 592.231390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.232061] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 592.239748] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.106s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.241318] env[61907]: INFO nova.compute.claims [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.752911] env[61907]: DEBUG nova.compute.utils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.755448] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 593.246198] env[61907]: INFO nova.scheduler.client.report [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Deleted allocations for instance dc04ac15-a97a-4248-b578-8f9046ea7e8e [ 593.262217] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 593.686733] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6095ad9a-e4ca-47be-ab1e-f8235a5b9288 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.702056] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42c70a0-92f3-4d0c-af6d-d8c4692bb782 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.743081] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad56ff79-ab87-4c17-953f-4b6ad93baee7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.752128] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68008b48-44e2-44b2-b496-33ead9df940e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.756407] env[61907]: DEBUG oslo_concurrency.lockutils [None req-71ee6b73-a93b-42b6-82b4-cbc53097f2e9 tempest-ImagesNegativeTestJSON-1260106327 tempest-ImagesNegativeTestJSON-1260106327-project-member] Lock "dc04ac15-a97a-4248-b578-8f9046ea7e8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.765s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.769324] env[61907]: DEBUG nova.compute.provider_tree [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.262083] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 594.274127] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 594.276952] env[61907]: DEBUG nova.scheduler.client.report [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 594.329867] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.330127] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.330286] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.330465] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.330601] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.330735] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.331547] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.331547] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.331547] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.331547] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.331932] env[61907]: DEBUG nova.virt.hardware [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.332767] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef372295-ee63-439c-9ce2-71e7c014a942 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.344027] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e9b316b-93b1-4dc1-8616-a0114070549c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.360718] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 594.373605] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.373605] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8270d143-cf3c-419c-8f87-a09bc2994519 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.388712] env[61907]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 594.388897] env[61907]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61907) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 594.389314] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Folder already exists: OpenStack. Parent ref: group-v4. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 594.389504] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating folder: Project (ba30b853723b4ce6be1c3dc7f3b24d9c). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.389751] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a59f6f0-41b1-4717-89ab-b77d14ed6749 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.403959] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Created folder: Project (ba30b853723b4ce6be1c3dc7f3b24d9c) in parent group-v268168. [ 594.404249] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating folder: Instances. Parent ref: group-v268177. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 594.405206] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-253a89cb-c656-48fe-98d3-4e8c8996b92e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.417818] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Created folder: Instances in parent group-v268177. [ 594.418035] env[61907]: DEBUG oslo.service.loopingcall [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 594.418195] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 594.418441] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a6b9ccce-7a5c-40f3-ae80-df437d4187b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.439681] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 594.439681] env[61907]: value = "task-1243592" [ 594.439681] env[61907]: _type = "Task" [ 594.439681] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.454967] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243592, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 594.519089] env[61907]: ERROR nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 594.519089] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.519089] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.519089] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.519089] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.519089] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.519089] env[61907]: ERROR nova.compute.manager raise self.value [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.519089] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.519089] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.519089] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.519703] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.519703] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.519703] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 594.519703] env[61907]: ERROR nova.compute.manager [ 594.519703] env[61907]: Traceback (most recent call last): [ 594.519703] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.519703] env[61907]: listener.cb(fileno) [ 594.519703] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.519703] env[61907]: result = function(*args, **kwargs) [ 594.519703] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.519703] env[61907]: return func(*args, **kwargs) [ 594.519703] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.519703] env[61907]: raise e [ 594.519703] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.519703] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 594.519703] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.519703] env[61907]: created_port_ids = self._update_ports_for_instance( [ 594.519703] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.519703] env[61907]: with excutils.save_and_reraise_exception(): [ 594.519703] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.519703] env[61907]: self.force_reraise() [ 594.519703] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.519703] env[61907]: raise self.value [ 594.519703] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.519703] env[61907]: updated_port = self._update_port( [ 594.519703] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.519703] env[61907]: _ensure_no_port_binding_failure(port) [ 594.519703] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.519703] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.520880] env[61907]: nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 594.520880] env[61907]: Removing descriptor: 15 [ 594.520880] env[61907]: ERROR nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Traceback (most recent call last): [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] yield resources [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.driver.spawn(context, instance, image_meta, [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.520880] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] vm_ref = self.build_virtual_machine(instance, [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] for vif in network_info: [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self._sync_wrapper(fn, *args, **kwargs) [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.wait() [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self[:] = self._gt.wait() [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self._exit_event.wait() [ 594.521260] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] result = hub.switch() [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self.greenlet.switch() [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] result = function(*args, **kwargs) [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return func(*args, **kwargs) [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise e [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] nwinfo = self.network_api.allocate_for_instance( [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.521645] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] created_port_ids = self._update_ports_for_instance( [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] with excutils.save_and_reraise_exception(): [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.force_reraise() [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise self.value [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] updated_port = self._update_port( [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] _ensure_no_port_binding_failure(port) [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.522164] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise exception.PortBindingFailed(port_id=port['id']) [ 594.522568] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 594.522568] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] [ 594.522568] env[61907]: INFO nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Terminating instance [ 594.788052] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.548s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 594.788493] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 594.797500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.445s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.802767] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.876253] env[61907]: DEBUG nova.compute.manager [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Received event network-changed-e1d0935d-af16-44a7-ad0e-1c8aa97c7074 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 594.876725] env[61907]: DEBUG nova.compute.manager [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Refreshing instance network info cache due to event network-changed-e1d0935d-af16-44a7-ad0e-1c8aa97c7074. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 594.876989] env[61907]: DEBUG oslo_concurrency.lockutils [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] Acquiring lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.877180] env[61907]: DEBUG oslo_concurrency.lockutils [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] Acquired lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.877368] env[61907]: DEBUG nova.network.neutron [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Refreshing network info cache for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 594.952279] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243592, 'name': CreateVM_Task, 'duration_secs': 0.283357} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 594.952461] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 594.953508] env[61907]: DEBUG oslo_vmware.service [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4bf0d2-5ec3-4db6-9c93-46d537d47412 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.960253] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.960444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.961101] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 594.961359] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22d77ddd-59f8-43cd-96c8-7e70e90baf22 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.966283] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 594.966283] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524bfc7c-0c07-0245-feeb-a8d30e2325db" [ 594.966283] env[61907]: _type = "Task" [ 594.966283] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 594.983480] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524bfc7c-0c07-0245-feeb-a8d30e2325db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 595.026866] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.306509] env[61907]: DEBUG nova.compute.utils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.307990] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 595.311695] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 595.473574] env[61907]: DEBUG nova.network.neutron [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.488732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.489013] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 595.489872] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.490903] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.490903] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 595.491084] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1fa8976-1fcb-45e4-97ae-d9c1bc30d064 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.501993] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 595.501993] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 595.503363] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd0d23b0-6707-4c13-8b2e-706862f95cf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.513521] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af76918b-57ce-4a69-be00-55837cffbe5e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.523666] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 595.523666] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52abc0a5-1752-128d-ffe1-8bb74843d6e8" [ 595.523666] env[61907]: _type = "Task" [ 595.523666] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 595.531029] env[61907]: DEBUG nova.policy [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb8984d341bb4fdf892048b6c9ec0be3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be91ad381ac848bfb21a4673d9fc8e77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.542127] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Preparing fetch location {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 595.542127] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating directory with path [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 595.542127] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f396a016-c199-4634-8300-d66b32d3f802 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.570931] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Created directory with path [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 595.571356] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Fetch image to [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 595.571634] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Downloading image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk on the data store datastore2 {{(pid=61907) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 595.572970] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d455f840-3dbc-40ee-9108-836ea1efc85b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.591250] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20138a1-fab0-4e82-9e31-f1fe21c40454 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.599160] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fa1afb-4bda-453d-99d3-1f0e76e12a99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.635605] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acc2cbf-edad-4ca0-b534-9f16c8c86e64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.652611] env[61907]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-52839318-f58f-4b83-8f59-9ab0cec82425 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.677128] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Downloading image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to the data store datastore2 {{(pid=61907) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 595.737170] env[61907]: DEBUG oslo_vmware.rw_handles [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61907) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 595.805047] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac3247b-bbe5-4bb5-8f24-377d1315187b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.814612] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 595.818098] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c962f0e2-5b69-4674-b16c-38719accdd2c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.856762] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304eb13b-fd54-4529-8dbd-fe75225a0911 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.865493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3ff584-72d4-444d-8153-7f113f1b185f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.880838] env[61907]: DEBUG nova.compute.provider_tree [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.088506] env[61907]: DEBUG nova.network.neutron [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.389991] env[61907]: DEBUG nova.scheduler.client.report [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 596.523832] env[61907]: DEBUG oslo_vmware.rw_handles [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Completed reading data from the image iterator. {{(pid=61907) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 596.524082] env[61907]: DEBUG oslo_vmware.rw_handles [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 596.589598] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Downloaded image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk on the data store datastore2 {{(pid=61907) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 596.592484] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Caching image {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 596.592484] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Copying Virtual Disk [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk to [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 596.592738] env[61907]: DEBUG oslo_concurrency.lockutils [req-035d4179-6d89-4bad-8a9c-0720d34c8216 req-7cb4413d-7917-474c-ada9-257e14aefbb8 service nova] Releasing lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.593617] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fc4ea1c7-173d-490d-befe-4e809ac7a93e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.595915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquired lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.596278] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.606710] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 596.606710] env[61907]: value = "task-1243593" [ 596.606710] env[61907]: _type = "Task" [ 596.606710] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 596.616306] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243593, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 596.831348] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 596.870695] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 596.874020] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 596.874020] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 596.874020] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 596.874020] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 596.874020] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 596.874289] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 596.874289] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 596.874289] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 596.874289] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 596.874289] env[61907]: DEBUG nova.virt.hardware [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 596.874450] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a5ae3f-b4b0-4aeb-a4b1-92765551f10e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.887198] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39ec6c22-6d4f-4674-a18f-48280a6048da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.908257] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.114s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.908967] env[61907]: ERROR nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Traceback (most recent call last): [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.driver.spawn(context, instance, image_meta, [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] vm_ref = self.build_virtual_machine(instance, [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.908967] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] for vif in network_info: [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self._sync_wrapper(fn, *args, **kwargs) [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.wait() [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self[:] = self._gt.wait() [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self._exit_event.wait() [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] result = hub.switch() [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.909388] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return self.greenlet.switch() [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] result = function(*args, **kwargs) [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] return func(*args, **kwargs) [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise e [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] nwinfo = self.network_api.allocate_for_instance( [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] created_port_ids = self._update_ports_for_instance( [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] with excutils.save_and_reraise_exception(): [ 596.909800] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] self.force_reraise() [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise self.value [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] updated_port = self._update_port( [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] _ensure_no_port_binding_failure(port) [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] raise exception.PortBindingFailed(port_id=port['id']) [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] nova.exception.PortBindingFailed: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. [ 596.910213] env[61907]: ERROR nova.compute.manager [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] [ 596.910599] env[61907]: DEBUG nova.compute.utils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 596.911938] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Build of instance 4db46e0e-1aaa-4168-8365-6dab93eee759 was re-scheduled: Binding failed for port 52d682ea-ac66-4d00-83f7-171ac8a91d72, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 596.913026] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 596.914551] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquiring lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.914551] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Acquired lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.914551] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.914859] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.085s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.917132] env[61907]: INFO nova.compute.claims [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.951952] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Successfully created port: a2f642bb-14c2-4767-8857-41f4dd1fe745 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.116149] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243593, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.157201] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.418842] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.500419] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "90a4f8bb-59bd-499f-b2fc-ebce55c5f803" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.500645] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "90a4f8bb-59bd-499f-b2fc-ebce55c5f803" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.508768] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.542766] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "6ac3cdf2-eb93-4403-a514-81d56baf544e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.543030] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "6ac3cdf2-eb93-4403-a514-81d56baf544e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.619205] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243593, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.702048} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 597.619205] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Copied Virtual Disk [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk to [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 597.619205] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleting the datastore file [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 597.619205] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e99bd187-e363-41ce-be15-fdfb4b2d727a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.630201] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 597.630201] env[61907]: value = "task-1243594" [ 597.630201] env[61907]: _type = "Task" [ 597.630201] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 597.642276] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 597.927401] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Releasing lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.928355] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 597.928355] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 597.931486] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-099e3147-5fbc-43e7-b138-a4acafb1f801 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.946768] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11ca5f1-afdb-43d1-ac11-187d2d17a6d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.978209] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0ec61517-a24f-4229-81e5-f889d1871ee7 could not be found. [ 597.978209] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 597.978438] env[61907]: INFO nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Took 0.05 seconds to destroy the instance on the hypervisor. [ 597.978538] env[61907]: DEBUG oslo.service.loopingcall [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 597.978741] env[61907]: DEBUG nova.compute.manager [-] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 597.978848] env[61907]: DEBUG nova.network.neutron [-] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.023733] env[61907]: DEBUG nova.network.neutron [-] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.106949] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.142276] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029134} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.142564] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 598.142769] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Moving file from [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57. {{(pid=61907) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 598.143035] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-8139fcf2-84da-4377-8ccc-4dc04a314319 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.155940] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 598.155940] env[61907]: value = "task-1243595" [ 598.155940] env[61907]: _type = "Task" [ 598.155940] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.169159] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243595, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.374965] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Successfully created port: 673e502d-4725-485f-98b7-c8b1fc0bca5b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 598.384874] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56614802-5b5d-4e16-886c-f75b3d43b85d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.393791] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a982582-edce-4fdb-8fa6-14a19cf5c17f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.425108] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c158b9-044e-4089-80d5-9522dd27d3cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.434300] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67ce585-7ce8-44e7-a73a-5759f44cda6a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.451783] env[61907]: DEBUG nova.compute.provider_tree [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 598.530396] env[61907]: DEBUG nova.network.neutron [-] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.616783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Releasing lock "refresh_cache-4db46e0e-1aaa-4168-8365-6dab93eee759" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.617288] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 598.617288] env[61907]: DEBUG nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 598.617450] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.666806] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243595, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.0345} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 598.667070] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] File moved {{(pid=61907) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 598.667266] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Cleaning up location [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 598.667428] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleting the datastore file [datastore2] vmware_temp/ca8f6aa7-6c85-4a00-8c7f-87abf404045e {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 598.667685] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c761777b-e4f5-40b7-873a-ea956c1f3332 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.675141] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 598.675141] env[61907]: value = "task-1243596" [ 598.675141] env[61907]: _type = "Task" [ 598.675141] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 598.685043] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 598.729292] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.759795] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "8456ccc3-aa32-4388-bf9e-2608c783405f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.760121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "8456ccc3-aa32-4388-bf9e-2608c783405f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.802898] env[61907]: DEBUG nova.compute.manager [req-47ab352e-1376-46bc-8b25-5134ba21ee20 req-ebf07aad-fa3c-458c-86b4-491409916dd4 service nova] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Received event network-vif-deleted-e1d0935d-af16-44a7-ad0e-1c8aa97c7074 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 598.954992] env[61907]: DEBUG nova.scheduler.client.report [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 599.038599] env[61907]: INFO nova.compute.manager [-] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Took 1.06 seconds to deallocate network for instance. [ 599.041502] env[61907]: DEBUG nova.compute.claims [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.041765] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.188995] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027086} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.189589] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 599.190395] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cce0d0e-f77e-4d74-b310-5bffd87bfef9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.196948] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 599.196948] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52223e55-08ca-88c7-7cfd-1442e8f7264f" [ 599.196948] env[61907]: _type = "Task" [ 599.196948] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.212207] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52223e55-08ca-88c7-7cfd-1442e8f7264f, 'name': SearchDatastore_Task, 'duration_secs': 0.009755} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.212452] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.212797] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 48b683f0-9cfe-472c-97df-1c19c3b5e441/48b683f0-9cfe-472c-97df-1c19c3b5e441.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 599.213168] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b99f39e5-6b90-4e0c-a52a-0da81d6bc50c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.220991] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 599.220991] env[61907]: value = "task-1243597" [ 599.220991] env[61907]: _type = "Task" [ 599.220991] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.237892] env[61907]: DEBUG nova.network.neutron [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.239332] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243597, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.460885] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.461612] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 599.465622] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.541s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.738197] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243597, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.508738} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 599.738510] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 48b683f0-9cfe-472c-97df-1c19c3b5e441/48b683f0-9cfe-472c-97df-1c19c3b5e441.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 599.740535] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 599.740535] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c991617-0d47-4786-8a78-7983e5bd81ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.741918] env[61907]: INFO nova.compute.manager [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] [instance: 4db46e0e-1aaa-4168-8365-6dab93eee759] Took 1.12 seconds to deallocate network for instance. [ 599.751700] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 599.751700] env[61907]: value = "task-1243598" [ 599.751700] env[61907]: _type = "Task" [ 599.751700] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 599.761070] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 599.975285] env[61907]: DEBUG nova.compute.utils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.986076] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 599.986076] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.064592] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Successfully created port: 149b4c1b-6429-4935-8cd9-d791862e3b18 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.264567] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068132} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.264944] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 600.265849] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247966c4-ab34-455f-b599-84b2f3fd7cb4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.273118] env[61907]: DEBUG nova.policy [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f564ee0cbe2044b69e4c5ff072d61cb3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a196dd116afe48fb976ed742db0d5f82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.301890] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Reconfiguring VM instance instance-0000000e to attach disk [datastore2] 48b683f0-9cfe-472c-97df-1c19c3b5e441/48b683f0-9cfe-472c-97df-1c19c3b5e441.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 600.305236] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c361b74a-e9ba-401e-be3d-a209e60e2116 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.326601] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 600.326601] env[61907]: value = "task-1243599" [ 600.326601] env[61907]: _type = "Task" [ 600.326601] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.334722] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 600.471730] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fac5f32-fecb-458d-b4d7-1070fa2453d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.479485] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428e2d96-444c-4263-97da-42d80ed1b1d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.486181] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 600.516425] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40322075-85dd-483f-9363-5c6882fc635d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.524374] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b78a2d3-0758-4fe0-91fe-259f190e6d77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.539971] env[61907]: DEBUG nova.compute.provider_tree [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.777401] env[61907]: INFO nova.scheduler.client.report [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Deleted allocations for instance 4db46e0e-1aaa-4168-8365-6dab93eee759 [ 600.837773] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243599, 'name': ReconfigVM_Task, 'duration_secs': 0.510961} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 600.838270] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Reconfigured VM instance instance-0000000e to attach disk [datastore2] 48b683f0-9cfe-472c-97df-1c19c3b5e441/48b683f0-9cfe-472c-97df-1c19c3b5e441.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 600.838894] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-afa3ec4e-656a-44fd-8e4a-edd0133a5c71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.845129] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 600.845129] env[61907]: value = "task-1243600" [ 600.845129] env[61907]: _type = "Task" [ 600.845129] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 600.855298] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243600, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.047167] env[61907]: DEBUG nova.scheduler.client.report [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 601.286026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd9275e4-346d-478e-af1c-5c89b6f2fa46 tempest-VolumesAssistedSnapshotsTest-583751824 tempest-VolumesAssistedSnapshotsTest-583751824-project-member] Lock "4db46e0e-1aaa-4168-8365-6dab93eee759" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.586s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.356926] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243600, 'name': Rename_Task, 'duration_secs': 0.136908} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.357723] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 601.357994] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa565586-f72b-4f74-9a0f-f9e608f759af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.364031] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 601.364031] env[61907]: value = "task-1243601" [ 601.364031] env[61907]: _type = "Task" [ 601.364031] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 601.371869] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243601, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 601.523580] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 601.552398] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 601.552663] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 601.552819] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 601.552996] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 601.553417] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 601.553585] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 601.553799] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 601.553960] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 601.554145] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 601.554313] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 601.554484] env[61907]: DEBUG nova.virt.hardware [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 601.555702] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.555948] env[61907]: ERROR nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Traceback (most recent call last): [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.driver.spawn(context, instance, image_meta, [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] vm_ref = self.build_virtual_machine(instance, [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.555948] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] for vif in network_info: [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self._sync_wrapper(fn, *args, **kwargs) [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.wait() [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self[:] = self._gt.wait() [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self._exit_event.wait() [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] result = hub.switch() [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.556463] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return self.greenlet.switch() [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] result = function(*args, **kwargs) [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] return func(*args, **kwargs) [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise e [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] nwinfo = self.network_api.allocate_for_instance( [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] created_port_ids = self._update_ports_for_instance( [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] with excutils.save_and_reraise_exception(): [ 601.556899] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] self.force_reraise() [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise self.value [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] updated_port = self._update_port( [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] _ensure_no_port_binding_failure(port) [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] raise exception.PortBindingFailed(port_id=port['id']) [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] nova.exception.PortBindingFailed: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. [ 601.557285] env[61907]: ERROR nova.compute.manager [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] [ 601.557613] env[61907]: DEBUG nova.compute.utils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.558864] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42ca597-2c0f-46ba-b77d-7ad18cfc6461 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.562959] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Build of instance 71013ca4-f1d0-4110-8fad-91e619ff354b was re-scheduled: Binding failed for port 7205218f-561b-4da0-9df1-4aa046a770aa, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 601.563200] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 601.563431] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquiring lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.563579] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Acquired lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.563737] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.564751] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.132s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.578835] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5352147-fe22-4e5a-acfe-2b55a3fd1288 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.789699] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 601.847907] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Successfully created port: 7a62a6ed-b052-4f40-ae09-f63fec67cd44 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.876442] env[61907]: DEBUG oslo_vmware.api [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243601, 'name': PowerOnVM_Task, 'duration_secs': 0.449185} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 601.876851] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 601.877157] env[61907]: INFO nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Took 7.60 seconds to spawn the instance on the hypervisor. [ 601.877420] env[61907]: DEBUG nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 601.878446] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5530097c-58c3-497f-8094-d0acb8a9542b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.121884] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.321494] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.406983] env[61907]: INFO nova.compute.manager [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Took 27.61 seconds to build instance. [ 602.412269] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbff8965-48c4-4e7b-8876-f15acbb9cc21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.420282] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d855ca-dc7b-4333-a835-17c9d826d959 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.424397] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.458340] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5279e02a-c40e-4c53-8823-b424ed73d6ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.467829] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c78b69-fdc1-46b7-b2f9-df9fcee201e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.483720] env[61907]: DEBUG nova.compute.provider_tree [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.910845] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6dba7d3f-58e7-47a2-a541-9cdc227b93db tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.635s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.927694] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Releasing lock "refresh_cache-71013ca4-f1d0-4110-8fad-91e619ff354b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.927694] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 602.927694] env[61907]: DEBUG nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 602.928636] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.970707] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.986477] env[61907]: DEBUG nova.scheduler.client.report [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 603.414922] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 603.477279] env[61907]: DEBUG nova.network.neutron [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.496084] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.931s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.496807] env[61907]: ERROR nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Traceback (most recent call last): [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.driver.spawn(context, instance, image_meta, [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] vm_ref = self.build_virtual_machine(instance, [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.496807] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] for vif in network_info: [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self._sync_wrapper(fn, *args, **kwargs) [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.wait() [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self[:] = self._gt.wait() [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self._exit_event.wait() [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] result = hub.switch() [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.498575] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return self.greenlet.switch() [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] result = function(*args, **kwargs) [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] return func(*args, **kwargs) [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise e [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] nwinfo = self.network_api.allocate_for_instance( [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] created_port_ids = self._update_ports_for_instance( [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] with excutils.save_and_reraise_exception(): [ 603.500373] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] self.force_reraise() [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise self.value [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] updated_port = self._update_port( [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] _ensure_no_port_binding_failure(port) [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] raise exception.PortBindingFailed(port_id=port['id']) [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] nova.exception.PortBindingFailed: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. [ 603.500992] env[61907]: ERROR nova.compute.manager [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] [ 603.501949] env[61907]: DEBUG nova.compute.utils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 603.501949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.469s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.502160] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Build of instance f66cffbb-1af4-4675-ae4c-ea6a034217c7 was re-scheduled: Binding failed for port b15d2a73-50b5-4506-9d63-68aeef31e1d0, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 603.502600] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 603.502836] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquiring lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.503060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Acquired lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.503685] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 603.952891] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.981710] env[61907]: INFO nova.compute.manager [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] [instance: 71013ca4-f1d0-4110-8fad-91e619ff354b] Took 1.05 seconds to deallocate network for instance. [ 604.044706] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.054427] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "520e2134-3553-44c2-98f5-ecd5be642d86" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.054702] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "520e2134-3553-44c2-98f5-ecd5be642d86" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.223804] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.405253] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e974a4cb-f4ac-4057-8990-ab661b84a87b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.415480] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb89dda3-ace0-4ec2-8d94-88539a9fe469 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.453193] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b7e1d3-6f5e-4d91-9e18-8b2c31bf5f12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.463123] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84d6255-d840-4d3b-a806-9774645fd32b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.479195] env[61907]: DEBUG nova.compute.provider_tree [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 604.727643] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Releasing lock "refresh_cache-f66cffbb-1af4-4675-ae4c-ea6a034217c7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.727643] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 604.727643] env[61907]: DEBUG nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 604.727643] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.766161] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.986436] env[61907]: DEBUG nova.scheduler.client.report [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 605.031738] env[61907]: INFO nova.scheduler.client.report [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Deleted allocations for instance 71013ca4-f1d0-4110-8fad-91e619ff354b [ 605.105497] env[61907]: ERROR nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 605.105497] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.105497] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.105497] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.105497] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.105497] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.105497] env[61907]: ERROR nova.compute.manager raise self.value [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.105497] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.105497] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.105497] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.106313] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.106313] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.106313] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 605.106313] env[61907]: ERROR nova.compute.manager [ 605.106313] env[61907]: Traceback (most recent call last): [ 605.106313] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.106313] env[61907]: listener.cb(fileno) [ 605.106313] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.106313] env[61907]: result = function(*args, **kwargs) [ 605.106313] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.106313] env[61907]: return func(*args, **kwargs) [ 605.106313] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.106313] env[61907]: raise e [ 605.106313] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.106313] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 605.106313] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.106313] env[61907]: created_port_ids = self._update_ports_for_instance( [ 605.106313] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.106313] env[61907]: with excutils.save_and_reraise_exception(): [ 605.106313] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.106313] env[61907]: self.force_reraise() [ 605.106313] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.106313] env[61907]: raise self.value [ 605.106313] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.106313] env[61907]: updated_port = self._update_port( [ 605.106313] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.106313] env[61907]: _ensure_no_port_binding_failure(port) [ 605.106313] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.106313] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.107747] env[61907]: nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 605.107747] env[61907]: Removing descriptor: 15 [ 605.107747] env[61907]: ERROR nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Traceback (most recent call last): [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] yield resources [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.driver.spawn(context, instance, image_meta, [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.107747] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] vm_ref = self.build_virtual_machine(instance, [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] for vif in network_info: [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self._sync_wrapper(fn, *args, **kwargs) [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.wait() [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self[:] = self._gt.wait() [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self._exit_event.wait() [ 605.108209] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] result = hub.switch() [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self.greenlet.switch() [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] result = function(*args, **kwargs) [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return func(*args, **kwargs) [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise e [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] nwinfo = self.network_api.allocate_for_instance( [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.108615] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] created_port_ids = self._update_ports_for_instance( [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] with excutils.save_and_reraise_exception(): [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.force_reraise() [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise self.value [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] updated_port = self._update_port( [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] _ensure_no_port_binding_failure(port) [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.112495] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise exception.PortBindingFailed(port_id=port['id']) [ 605.112909] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 605.112909] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] [ 605.112909] env[61907]: INFO nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Terminating instance [ 605.210049] env[61907]: DEBUG nova.compute.manager [None req-a80c9bad-292e-414d-b0e4-3a32bd001b2a tempest-ServerDiagnosticsV248Test-1333224293 tempest-ServerDiagnosticsV248Test-1333224293-project-admin] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 605.211434] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16485948-0764-455f-a113-785b6ca78251 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.220826] env[61907]: INFO nova.compute.manager [None req-a80c9bad-292e-414d-b0e4-3a32bd001b2a tempest-ServerDiagnosticsV248Test-1333224293 tempest-ServerDiagnosticsV248Test-1333224293-project-admin] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Retrieving diagnostics [ 605.221524] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5151cb48-ebff-45a4-84ef-4d20b1c686a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.269497] env[61907]: DEBUG nova.network.neutron [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.324262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquiring lock "a1189e41-5922-4623-bb03-394d98159193" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.324262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "a1189e41-5922-4623-bb03-394d98159193" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.490674] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.491510] env[61907]: ERROR nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Traceback (most recent call last): [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.driver.spawn(context, instance, image_meta, [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] vm_ref = self.build_virtual_machine(instance, [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.491510] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] for vif in network_info: [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self._sync_wrapper(fn, *args, **kwargs) [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.wait() [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self[:] = self._gt.wait() [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self._exit_event.wait() [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] result = hub.switch() [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.491923] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return self.greenlet.switch() [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] result = function(*args, **kwargs) [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] return func(*args, **kwargs) [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise e [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] nwinfo = self.network_api.allocate_for_instance( [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] created_port_ids = self._update_ports_for_instance( [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] with excutils.save_and_reraise_exception(): [ 605.492324] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] self.force_reraise() [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise self.value [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] updated_port = self._update_port( [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] _ensure_no_port_binding_failure(port) [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] raise exception.PortBindingFailed(port_id=port['id']) [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] nova.exception.PortBindingFailed: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. [ 605.492716] env[61907]: ERROR nova.compute.manager [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] [ 605.496713] env[61907]: DEBUG nova.compute.utils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 605.496713] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.212s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.498245] env[61907]: INFO nova.compute.claims [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 605.501183] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Build of instance b3937fb5-74f4-4832-b9cf-9c51ee8ceb50 was re-scheduled: Binding failed for port df82cc2c-0312-46b4-813a-318ad3ff342d, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 605.501683] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 605.501874] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.502048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquired lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.502210] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.538218] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45e0028a-50fa-4d37-b917-8c559c9d02f2 tempest-ServersWithSpecificFlavorTestJSON-792447156 tempest-ServersWithSpecificFlavorTestJSON-792447156-project-member] Lock "71013ca4-f1d0-4110-8fad-91e619ff354b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.046s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.607691] env[61907]: ERROR nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 605.607691] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.607691] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.607691] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.607691] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.607691] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.607691] env[61907]: ERROR nova.compute.manager raise self.value [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.607691] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 605.607691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.607691] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 605.608244] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.608244] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 605.608244] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 605.608244] env[61907]: ERROR nova.compute.manager [ 605.608244] env[61907]: Traceback (most recent call last): [ 605.608244] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 605.608244] env[61907]: listener.cb(fileno) [ 605.608244] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.608244] env[61907]: result = function(*args, **kwargs) [ 605.608244] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.608244] env[61907]: return func(*args, **kwargs) [ 605.608244] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.608244] env[61907]: raise e [ 605.608244] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.608244] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 605.608244] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.608244] env[61907]: created_port_ids = self._update_ports_for_instance( [ 605.608244] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.608244] env[61907]: with excutils.save_and_reraise_exception(): [ 605.608244] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.608244] env[61907]: self.force_reraise() [ 605.608244] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.608244] env[61907]: raise self.value [ 605.608244] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.608244] env[61907]: updated_port = self._update_port( [ 605.608244] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.608244] env[61907]: _ensure_no_port_binding_failure(port) [ 605.608244] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.608244] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 605.609122] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 605.609122] env[61907]: Removing descriptor: 17 [ 605.609122] env[61907]: ERROR nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Traceback (most recent call last): [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] yield resources [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.driver.spawn(context, instance, image_meta, [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 605.609122] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] vm_ref = self.build_virtual_machine(instance, [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] vif_infos = vmwarevif.get_vif_info(self._session, [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] for vif in network_info: [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self._sync_wrapper(fn, *args, **kwargs) [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.wait() [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self[:] = self._gt.wait() [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self._exit_event.wait() [ 605.609523] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] result = hub.switch() [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self.greenlet.switch() [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] result = function(*args, **kwargs) [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return func(*args, **kwargs) [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise e [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] nwinfo = self.network_api.allocate_for_instance( [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 605.609958] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] created_port_ids = self._update_ports_for_instance( [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] with excutils.save_and_reraise_exception(): [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.force_reraise() [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise self.value [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] updated_port = self._update_port( [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] _ensure_no_port_binding_failure(port) [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 605.610378] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise exception.PortBindingFailed(port_id=port['id']) [ 605.610759] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 605.610759] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] [ 605.610759] env[61907]: INFO nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Terminating instance [ 605.617129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 605.617129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquired lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 605.617824] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 605.777295] env[61907]: INFO nova.compute.manager [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] [instance: f66cffbb-1af4-4675-ae4c-ea6a034217c7] Took 1.05 seconds to deallocate network for instance. [ 606.028337] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.042209] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 606.115495] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquiring lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.115685] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquired lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.116232] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 606.117658] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.192199] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.374055] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.379459] env[61907]: DEBUG nova.compute.manager [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Received event network-changed-a2f642bb-14c2-4767-8857-41f4dd1fe745 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 606.379459] env[61907]: DEBUG nova.compute.manager [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Refreshing instance network info cache due to event network-changed-a2f642bb-14c2-4767-8857-41f4dd1fe745. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 606.379459] env[61907]: DEBUG oslo_concurrency.lockutils [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] Acquiring lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.565958] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.624413] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Releasing lock "refresh_cache-b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.624413] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 606.624413] env[61907]: DEBUG nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 606.624413] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 606.673182] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.674503] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.845054] env[61907]: INFO nova.scheduler.client.report [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Deleted allocations for instance f66cffbb-1af4-4675-ae4c-ea6a034217c7 [ 606.862026] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.869430] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquiring lock "a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.869560] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.880064] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Releasing lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 606.880489] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 606.880673] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 606.880966] env[61907]: DEBUG oslo_concurrency.lockutils [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] Acquired lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.881159] env[61907]: DEBUG nova.network.neutron [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Refreshing network info cache for port a2f642bb-14c2-4767-8857-41f4dd1fe745 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.882208] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15d52e4b-ac52-4b60-afc1-0ad818f76c7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.898024] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67b79ea-e223-4da2-b07a-019f37008d29 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.924951] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5556aefa-3c0b-415c-94f3-603c7fb26ec1 could not be found. [ 606.925245] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 606.925433] env[61907]: INFO nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 606.929890] env[61907]: DEBUG oslo.service.loopingcall [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 606.933971] env[61907]: DEBUG nova.compute.manager [-] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 606.933971] env[61907]: DEBUG nova.network.neutron [-] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.004020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd51d1e1-2a8f-49ca-840b-a3b1ca8d3843 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.012593] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a904376d-473d-4851-b891-cbf46de462bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.050840] env[61907]: DEBUG nova.network.neutron [-] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.053776] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f528e60-e07f-411c-bf5f-049e85b81e0e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.064090] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc80293b-5e13-4f62-ab12-32afb57c83be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.078967] env[61907]: DEBUG nova.compute.provider_tree [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.179817] env[61907]: DEBUG nova.network.neutron [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.375876] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9ab2325e-670d-4e05-bd34-a6b2a76db34a tempest-ServersAdminNegativeTestJSON-1769547351 tempest-ServersAdminNegativeTestJSON-1769547351-project-member] Lock "f66cffbb-1af4-4675-ae4c-ea6a034217c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.349s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.379134] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Releasing lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.379134] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 607.379134] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 607.385880] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04638c4b-82a7-4853-a028-b4303ea997b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.394116] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquiring lock "1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.394116] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.403523] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c6898e-e5fc-4fba-aa38-8e34183bfc03 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.427502] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance bed89649-b8d3-414a-8729-a504a7e02b94 could not be found. [ 607.427727] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 607.427905] env[61907]: INFO nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Took 0.05 seconds to destroy the instance on the hypervisor. [ 607.428219] env[61907]: DEBUG oslo.service.loopingcall [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 607.428463] env[61907]: DEBUG nova.compute.manager [-] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 607.428558] env[61907]: DEBUG nova.network.neutron [-] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 607.453833] env[61907]: DEBUG nova.network.neutron [-] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.455875] env[61907]: DEBUG nova.network.neutron [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 607.582864] env[61907]: DEBUG nova.scheduler.client.report [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 607.626731] env[61907]: DEBUG nova.network.neutron [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.683865] env[61907]: INFO nova.compute.manager [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: b3937fb5-74f4-4832-b9cf-9c51ee8ceb50] Took 1.06 seconds to deallocate network for instance. [ 607.893305] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 607.958527] env[61907]: DEBUG nova.network.neutron [-] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.091528] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.595s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.092116] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 608.095587] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.293s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.099916] env[61907]: INFO nova.compute.claims [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.131585] env[61907]: DEBUG oslo_concurrency.lockutils [req-eff0a9d0-b987-401b-bac2-ef98ef7f8c94 req-698b3f00-40df-409c-a69c-0199e4b388f8 service nova] Releasing lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.266711] env[61907]: DEBUG nova.network.neutron [-] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.307500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquiring lock "81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.307852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.420787] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.462714] env[61907]: INFO nova.compute.manager [-] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Took 1.03 seconds to deallocate network for instance. [ 608.465267] env[61907]: DEBUG nova.compute.claims [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.465453] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.606412] env[61907]: DEBUG nova.compute.utils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 608.607866] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 608.608010] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 608.720866] env[61907]: DEBUG nova.policy [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '772dbe17f5cd4dc9904197e708e3f316', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3aab40034a4e4ccdb169abfc11f3b02f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.732037] env[61907]: INFO nova.scheduler.client.report [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Deleted allocations for instance b3937fb5-74f4-4832-b9cf-9c51ee8ceb50 [ 608.770485] env[61907]: INFO nova.compute.manager [-] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Took 1.84 seconds to deallocate network for instance. [ 608.771352] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.771352] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.773501] env[61907]: DEBUG nova.compute.claims [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 608.773664] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.088994] env[61907]: DEBUG nova.compute.manager [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Received event network-vif-deleted-a2f642bb-14c2-4767-8857-41f4dd1fe745 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 609.089239] env[61907]: DEBUG nova.compute.manager [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Received event network-changed-7a62a6ed-b052-4f40-ae09-f63fec67cd44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 609.089406] env[61907]: DEBUG nova.compute.manager [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Refreshing instance network info cache due to event network-changed-7a62a6ed-b052-4f40-ae09-f63fec67cd44. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 609.089615] env[61907]: DEBUG oslo_concurrency.lockutils [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] Acquiring lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.089753] env[61907]: DEBUG oslo_concurrency.lockutils [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] Acquired lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.089908] env[61907]: DEBUG nova.network.neutron [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Refreshing network info cache for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 609.122454] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 609.248322] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61b35f82-d6bb-48d1-b110-a8034e56b193 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "b3937fb5-74f4-4832-b9cf-9c51ee8ceb50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.487s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.320099] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Successfully created port: 8885367d-c594-44d9-9ae6-13c3f30bee98 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 609.598740] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b937061d-85e8-4d59-a00c-490d36786372 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.607176] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-810b36f7-f2d8-4a6d-9af6-2073e46cb2c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.657700] env[61907]: DEBUG nova.network.neutron [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.664014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c9149c9-de80-469c-9c09-68c0acda630e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.673023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebc8c62-2663-42c7-8157-bed5ddd6cd6b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.685595] env[61907]: DEBUG nova.compute.provider_tree [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.751090] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 609.932199] env[61907]: DEBUG nova.network.neutron [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.158839] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 610.191196] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 610.191196] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 610.191196] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 610.191557] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 610.192583] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 610.194650] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 610.194650] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 610.194650] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 610.194650] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 610.194650] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 610.195128] env[61907]: DEBUG nova.virt.hardware [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 610.195128] env[61907]: DEBUG nova.scheduler.client.report [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 610.200584] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c9e17f-2462-4b04-8f41-c9d8c8006dce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.214081] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd95456-6819-48bd-8242-de57b55b2911 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.282231] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.435648] env[61907]: DEBUG oslo_concurrency.lockutils [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] Releasing lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.436180] env[61907]: DEBUG nova.compute.manager [req-2608d67c-6e10-48ab-9406-fad2ea2b1c48 req-068da31d-f613-464f-9503-baf893929095 service nova] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Received event network-vif-deleted-7a62a6ed-b052-4f40-ae09-f63fec67cd44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 610.701191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.605s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.702039] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 610.704011] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.662s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.212247] env[61907]: DEBUG nova.compute.utils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.215317] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 611.216406] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 611.326172] env[61907]: DEBUG nova.policy [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62cc64e08b024d6190f003524f9c7797', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8f34803a44ee4eeda576105f6eb425e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 611.479835] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "09d97063-d94f-439a-a811-b3039de1f4d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.480826] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "09d97063-d94f-439a-a811-b3039de1f4d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.659305] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32b83a8-2b00-4abd-96e6-0fcea0ed5bfd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.666670] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f3c4ec-6774-4bfe-8951-e1f9c3682bf3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.701670] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb97e19d-0257-4872-b68c-2a8ba07a65c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.711213] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d17a5ad-1f9d-482a-a905-0e6206b819f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.716588] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 611.729892] env[61907]: DEBUG nova.compute.provider_tree [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.835216] env[61907]: DEBUG nova.compute.manager [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Received event network-changed-8885367d-c594-44d9-9ae6-13c3f30bee98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 611.835412] env[61907]: DEBUG nova.compute.manager [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Refreshing instance network info cache due to event network-changed-8885367d-c594-44d9-9ae6-13c3f30bee98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 611.835707] env[61907]: DEBUG oslo_concurrency.lockutils [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] Acquiring lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.835873] env[61907]: DEBUG oslo_concurrency.lockutils [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] Acquired lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.836056] env[61907]: DEBUG nova.network.neutron [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Refreshing network info cache for port 8885367d-c594-44d9-9ae6-13c3f30bee98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 612.233399] env[61907]: DEBUG nova.scheduler.client.report [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 612.347969] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Successfully created port: 4349e1bc-b7ad-4ceb-8e38-276020ab0189 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.386845] env[61907]: DEBUG nova.network.neutron [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.404326] env[61907]: ERROR nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 612.404326] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.404326] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.404326] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.404326] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.404326] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.404326] env[61907]: ERROR nova.compute.manager raise self.value [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.404326] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 612.404326] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.404326] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 612.404700] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.404700] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 612.404700] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 612.404700] env[61907]: ERROR nova.compute.manager [ 612.404700] env[61907]: Traceback (most recent call last): [ 612.404700] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 612.404700] env[61907]: listener.cb(fileno) [ 612.404700] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.404700] env[61907]: result = function(*args, **kwargs) [ 612.404700] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.404700] env[61907]: return func(*args, **kwargs) [ 612.404700] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.404700] env[61907]: raise e [ 612.404700] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.404700] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 612.404700] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.404700] env[61907]: created_port_ids = self._update_ports_for_instance( [ 612.404700] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.404700] env[61907]: with excutils.save_and_reraise_exception(): [ 612.404700] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.404700] env[61907]: self.force_reraise() [ 612.404700] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.404700] env[61907]: raise self.value [ 612.404700] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.404700] env[61907]: updated_port = self._update_port( [ 612.404700] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.404700] env[61907]: _ensure_no_port_binding_failure(port) [ 612.404700] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.404700] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 612.405260] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 612.405260] env[61907]: Removing descriptor: 17 [ 612.405260] env[61907]: ERROR nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Traceback (most recent call last): [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] yield resources [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.driver.spawn(context, instance, image_meta, [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.405260] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] vm_ref = self.build_virtual_machine(instance, [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] for vif in network_info: [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self._sync_wrapper(fn, *args, **kwargs) [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.wait() [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self[:] = self._gt.wait() [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self._exit_event.wait() [ 612.407853] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] result = hub.switch() [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self.greenlet.switch() [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] result = function(*args, **kwargs) [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return func(*args, **kwargs) [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise e [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] nwinfo = self.network_api.allocate_for_instance( [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.408137] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] created_port_ids = self._update_ports_for_instance( [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] with excutils.save_and_reraise_exception(): [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.force_reraise() [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise self.value [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] updated_port = self._update_port( [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] _ensure_no_port_binding_failure(port) [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.408454] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise exception.PortBindingFailed(port_id=port['id']) [ 612.408703] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 612.408703] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] [ 612.408703] env[61907]: INFO nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Terminating instance [ 612.551382] env[61907]: DEBUG nova.network.neutron [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.728731] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 612.745199] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.041s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.746832] env[61907]: ERROR nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Traceback (most recent call last): [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.driver.spawn(context, instance, image_meta, [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] vm_ref = self.build_virtual_machine(instance, [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.746832] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] for vif in network_info: [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self._sync_wrapper(fn, *args, **kwargs) [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.wait() [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self[:] = self._gt.wait() [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self._exit_event.wait() [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] result = hub.switch() [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.747166] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return self.greenlet.switch() [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] result = function(*args, **kwargs) [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] return func(*args, **kwargs) [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise e [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] nwinfo = self.network_api.allocate_for_instance( [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] created_port_ids = self._update_ports_for_instance( [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] with excutils.save_and_reraise_exception(): [ 612.747452] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] self.force_reraise() [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise self.value [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] updated_port = self._update_port( [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] _ensure_no_port_binding_failure(port) [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] raise exception.PortBindingFailed(port_id=port['id']) [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] nova.exception.PortBindingFailed: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. [ 612.747730] env[61907]: ERROR nova.compute.manager [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] [ 612.747966] env[61907]: DEBUG nova.compute.utils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.748953] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.428s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.750767] env[61907]: INFO nova.compute.claims [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.755829] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Build of instance 0ec61517-a24f-4229-81e5-f889d1871ee7 was re-scheduled: Binding failed for port e1d0935d-af16-44a7-ad0e-1c8aa97c7074, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 612.755829] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 612.755829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquiring lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.755829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Acquired lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.755948] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.761429] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 612.761661] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 612.761819] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 612.761998] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 612.762164] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 612.762311] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 612.762512] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 612.762667] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 612.762828] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 612.762986] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 612.763173] env[61907]: DEBUG nova.virt.hardware [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.764039] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-437ebc6f-d06f-4ce4-8b38-f0d6c39ae466 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.773255] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e859a591-bb82-4fb6-8eca-668622532416 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.913249] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.937699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "af575382-25f6-475a-8803-48714fa0a37c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.937934] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "af575382-25f6-475a-8803-48714fa0a37c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.055709] env[61907]: DEBUG oslo_concurrency.lockutils [req-1bdca918-5144-45fc-a792-77d1fbecf457 req-98720719-182a-4530-9364-8fbeb91e1f21 service nova] Releasing lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.055709] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.055709] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.322047] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.399403] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.594704] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.861011] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.900775] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Releasing lock "refresh_cache-0ec61517-a24f-4229-81e5-f889d1871ee7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.900775] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 613.900775] env[61907]: DEBUG nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 613.901579] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.938274] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.975745] env[61907]: DEBUG nova.compute.manager [req-012e82fb-86d4-4e34-839c-8b71bb81843a req-1861c031-4e02-47bd-acf8-4ee8b0d77d86 service nova] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Received event network-vif-deleted-8885367d-c594-44d9-9ae6-13c3f30bee98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 614.189207] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fdda20-d38d-41fe-a7be-bf655519da43 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.198278] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24d53f69-c7d0-4083-8a2b-c9b788d0e130 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.233348] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5647e6-6b6b-430e-beee-3783c9275928 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.241824] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7f224e-d42b-443d-a4be-15fd10ba38a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.256176] env[61907]: DEBUG nova.compute.provider_tree [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.368178] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.370427] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 614.371100] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.371100] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03e75083-e6f2-42a2-93d1-b326016d44c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.382930] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca48897-05ed-4fe9-bb08-43f58b18a920 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.414915] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c6fd5d45-1abd-4a12-8fcf-86930161bafb could not be found. [ 614.415123] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.415294] env[61907]: INFO nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.415535] env[61907]: DEBUG oslo.service.loopingcall [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.415768] env[61907]: DEBUG nova.compute.manager [-] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 614.415870] env[61907]: DEBUG nova.network.neutron [-] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.440614] env[61907]: DEBUG nova.network.neutron [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.450162] env[61907]: DEBUG nova.network.neutron [-] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.760238] env[61907]: DEBUG nova.scheduler.client.report [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 614.945331] env[61907]: INFO nova.compute.manager [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] [instance: 0ec61517-a24f-4229-81e5-f889d1871ee7] Took 1.04 seconds to deallocate network for instance. [ 614.953090] env[61907]: DEBUG nova.network.neutron [-] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.269722] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.270127] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 615.274626] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.322s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.276962] env[61907]: INFO nova.compute.claims [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.459841] env[61907]: INFO nova.compute.manager [-] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Took 1.04 seconds to deallocate network for instance. [ 615.463925] env[61907]: DEBUG nova.compute.claims [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.464570] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.691141] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.691430] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.776035] env[61907]: DEBUG nova.compute.utils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.778977] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 615.778977] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.781753] env[61907]: ERROR nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 615.781753] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.781753] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.781753] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.781753] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.781753] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.781753] env[61907]: ERROR nova.compute.manager raise self.value [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.781753] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 615.781753] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.781753] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 615.782144] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.782144] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 615.782144] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 615.782144] env[61907]: ERROR nova.compute.manager [ 615.782144] env[61907]: Traceback (most recent call last): [ 615.782144] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 615.782144] env[61907]: listener.cb(fileno) [ 615.782144] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.782144] env[61907]: result = function(*args, **kwargs) [ 615.782144] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.782144] env[61907]: return func(*args, **kwargs) [ 615.782144] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 615.782144] env[61907]: raise e [ 615.782144] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.782144] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 615.782144] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.782144] env[61907]: created_port_ids = self._update_ports_for_instance( [ 615.782144] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.782144] env[61907]: with excutils.save_and_reraise_exception(): [ 615.782144] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.782144] env[61907]: self.force_reraise() [ 615.782144] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.782144] env[61907]: raise self.value [ 615.782144] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.782144] env[61907]: updated_port = self._update_port( [ 615.782144] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.782144] env[61907]: _ensure_no_port_binding_failure(port) [ 615.782144] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.782144] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 615.782826] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 615.782826] env[61907]: Removing descriptor: 15 [ 615.784756] env[61907]: ERROR nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Traceback (most recent call last): [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] yield resources [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.driver.spawn(context, instance, image_meta, [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] vm_ref = self.build_virtual_machine(instance, [ 615.784756] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] for vif in network_info: [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self._sync_wrapper(fn, *args, **kwargs) [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.wait() [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self[:] = self._gt.wait() [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self._exit_event.wait() [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.785172] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] result = hub.switch() [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self.greenlet.switch() [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] result = function(*args, **kwargs) [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return func(*args, **kwargs) [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise e [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] nwinfo = self.network_api.allocate_for_instance( [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] created_port_ids = self._update_ports_for_instance( [ 615.785436] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] with excutils.save_and_reraise_exception(): [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.force_reraise() [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise self.value [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] updated_port = self._update_port( [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] _ensure_no_port_binding_failure(port) [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise exception.PortBindingFailed(port_id=port['id']) [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 615.785704] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] [ 615.785967] env[61907]: INFO nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Terminating instance [ 615.868330] env[61907]: DEBUG nova.policy [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6cf54ef44058461aa1e1d7be79959f21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '21eaf66fc0d24a7db499af006deac133', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.992411] env[61907]: INFO nova.scheduler.client.report [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Deleted allocations for instance 0ec61517-a24f-4229-81e5-f889d1871ee7 [ 616.142061] env[61907]: DEBUG nova.compute.manager [None req-8259107f-cef2-40f9-a150-034ee13d510c tempest-ServerDiagnosticsV248Test-1333224293 tempest-ServerDiagnosticsV248Test-1333224293-project-admin] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 616.145657] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87abc903-36ed-4707-aa57-ce9438e5b8da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.159865] env[61907]: INFO nova.compute.manager [None req-8259107f-cef2-40f9-a150-034ee13d510c tempest-ServerDiagnosticsV248Test-1333224293 tempest-ServerDiagnosticsV248Test-1333224293-project-admin] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Retrieving diagnostics [ 616.163043] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-446d2fcf-d1cc-4a95-a891-f548d1fb198b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.283285] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 616.299797] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquiring lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.299797] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquired lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.299797] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 616.499747] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Successfully created port: a4de15f4-98c4-4297-8476-461ece4ed7a2 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.510932] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a6fccaef-1c9e-424e-bcd3-7b1a155b5a12 tempest-DeleteServersAdminTestJSON-57113111 tempest-DeleteServersAdminTestJSON-57113111-project-member] Lock "0ec61517-a24f-4229-81e5-f889d1871ee7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.239s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.691651] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.691883] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.727773] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cfae27-b796-4c02-a3ed-1c7120f38ab2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.736625] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3afef48d-7e90-4401-a8f3-228af44d2072 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.775700] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5421fe-decf-402f-9133-a18aa6e5724f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.780511] env[61907]: DEBUG nova.compute.manager [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Received event network-changed-4349e1bc-b7ad-4ceb-8e38-276020ab0189 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 616.780712] env[61907]: DEBUG nova.compute.manager [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Refreshing instance network info cache due to event network-changed-4349e1bc-b7ad-4ceb-8e38-276020ab0189. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 616.780965] env[61907]: DEBUG oslo_concurrency.lockutils [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] Acquiring lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 616.787909] env[61907]: INFO nova.virt.block_device [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Booting with volume 654453e7-2370-4652-9285-d69da695b9c4 at /dev/sda [ 616.791919] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086c801e-5ec7-4bfa-b39c-d4dc3202e6ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.809075] env[61907]: DEBUG nova.compute.provider_tree [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.821828] env[61907]: DEBUG nova.scheduler.client.report [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 616.847065] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.858263] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-141a5fef-bccb-4a32-84bc-cb9100f193bf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.867359] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d56f99d-17fb-448c-9274-cd4df7e9094c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.889815] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a1e84fa-c229-4cc8-b3c0-8823780be26d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.897184] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671320a5-5dbb-456d-83f8-39594f215f30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.920020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5298215a-b177-4d1d-9bd7-82854063389d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.924953] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183903e6-40ab-463b-834c-c1aff101917d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.939217] env[61907]: DEBUG nova.virt.block_device [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating existing volume attachment record: 4d7fd89f-0c39-4a01-a2d6-85f634cb3943 {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 617.017851] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 617.029033] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.187188] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquiring lock "4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.187188] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.329377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.053s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.329377] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 617.334164] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.768s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.336084] env[61907]: INFO nova.compute.claims [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.531281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Releasing lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.533274] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 617.533274] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 617.533274] env[61907]: DEBUG oslo_concurrency.lockutils [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] Acquired lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.533274] env[61907]: DEBUG nova.network.neutron [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Refreshing network info cache for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.533689] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e35cee1-1825-4ef5-8ba2-5ac87986f2d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.544388] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fdf77ac-716f-4db7-9a30-120fa1a8b51f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.559088] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.576014] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2fcb4164-c0c5-49e5-87cb-a51df711b5a2 could not be found. [ 617.576014] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 617.576014] env[61907]: INFO nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 617.576240] env[61907]: DEBUG oslo.service.loopingcall [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 617.576765] env[61907]: DEBUG nova.compute.manager [-] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 617.576964] env[61907]: DEBUG nova.network.neutron [-] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 617.600535] env[61907]: DEBUG nova.network.neutron [-] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 617.774591] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "ea6d2b19-be25-4da3-b8af-11c3dca04df8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.775183] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "ea6d2b19-be25-4da3-b8af-11c3dca04df8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.841736] env[61907]: DEBUG nova.compute.utils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.845395] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 617.848759] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.947807] env[61907]: DEBUG nova.policy [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12e6e3d6b4d54da1b39b1b9ba210d4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1e1197870d94715bf949e49ca6ce9f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 618.062420] env[61907]: DEBUG nova.network.neutron [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.104187] env[61907]: DEBUG nova.network.neutron [-] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.226947] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.227287] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.229019] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "48b683f0-9cfe-472c-97df-1c19c3b5e441-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.229019] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.229019] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.231724] env[61907]: INFO nova.compute.manager [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Terminating instance [ 618.349561] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 618.399662] env[61907]: DEBUG nova.compute.manager [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Received event network-changed-a4de15f4-98c4-4297-8476-461ece4ed7a2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 618.399848] env[61907]: DEBUG nova.compute.manager [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Refreshing instance network info cache due to event network-changed-a4de15f4-98c4-4297-8476-461ece4ed7a2. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 618.400065] env[61907]: DEBUG oslo_concurrency.lockutils [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] Acquiring lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.400370] env[61907]: DEBUG oslo_concurrency.lockutils [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] Acquired lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.400562] env[61907]: DEBUG nova.network.neutron [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Refreshing network info cache for port a4de15f4-98c4-4297-8476-461ece4ed7a2 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 618.410348] env[61907]: ERROR nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 618.410348] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.410348] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.410348] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.410348] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.410348] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.410348] env[61907]: ERROR nova.compute.manager raise self.value [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.410348] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 618.410348] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.410348] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 618.410832] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.410832] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 618.410832] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 618.410832] env[61907]: ERROR nova.compute.manager [ 618.410832] env[61907]: Traceback (most recent call last): [ 618.410832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 618.410832] env[61907]: listener.cb(fileno) [ 618.410832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.410832] env[61907]: result = function(*args, **kwargs) [ 618.410832] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.410832] env[61907]: return func(*args, **kwargs) [ 618.410832] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 618.410832] env[61907]: raise e [ 618.410832] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 618.410832] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 618.410832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 618.410832] env[61907]: created_port_ids = self._update_ports_for_instance( [ 618.410832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 618.410832] env[61907]: with excutils.save_and_reraise_exception(): [ 618.410832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.410832] env[61907]: self.force_reraise() [ 618.410832] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.410832] env[61907]: raise self.value [ 618.410832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 618.410832] env[61907]: updated_port = self._update_port( [ 618.410832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.410832] env[61907]: _ensure_no_port_binding_failure(port) [ 618.410832] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.410832] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 618.414174] env[61907]: nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 618.414174] env[61907]: Removing descriptor: 18 [ 618.499244] env[61907]: DEBUG nova.network.neutron [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.610511] env[61907]: INFO nova.compute.manager [-] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Took 1.03 seconds to deallocate network for instance. [ 618.613839] env[61907]: DEBUG nova.compute.claims [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 618.614033] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.740511] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.740729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquired lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.740906] env[61907]: DEBUG nova.network.neutron [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 618.799101] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Successfully created port: 20e74284-c6f7-4668-bdbd-161482afcbcb {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.848040] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0602b422-0b9a-4896-ad83-ee77eb0cb9d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.854780] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd9018a-66e6-447e-a188-bc382539e8d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.894548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50cede6-47c4-4ac8-ab57-fb57761bc91c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.903483] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db14246f-a70b-4ca1-b8be-c96cf2d11f88 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.922182] env[61907]: DEBUG nova.compute.provider_tree [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.961024] env[61907]: DEBUG nova.network.neutron [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.005331] env[61907]: DEBUG oslo_concurrency.lockutils [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] Releasing lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.006032] env[61907]: DEBUG nova.compute.manager [req-482cb0bb-a992-4ce8-9d36-63920b24b840 req-e4a47a51-d3c1-4ca5-a9c6-b398e22ba1e0 service nova] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Received event network-vif-deleted-4349e1bc-b7ad-4ceb-8e38-276020ab0189 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 619.054826] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 619.055634] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.056763] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.056973] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.057371] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.057371] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.057648] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.057829] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.058060] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.058945] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.058945] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.058945] env[61907]: DEBUG nova.virt.hardware [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.059759] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee6b8806-f8b3-4e32-b09d-9cf9cf4704dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.069142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3a4f385-258b-4516-b8f5-07196f261471 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.085995] env[61907]: ERROR nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Traceback (most recent call last): [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] yield resources [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.driver.spawn(context, instance, image_meta, [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] vm_ref = self.build_virtual_machine(instance, [ 619.085995] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] for vif in network_info: [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return self._sync_wrapper(fn, *args, **kwargs) [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.wait() [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self[:] = self._gt.wait() [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return self._exit_event.wait() [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 619.086310] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] current.throw(*self._exc) [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] result = function(*args, **kwargs) [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return func(*args, **kwargs) [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise e [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] nwinfo = self.network_api.allocate_for_instance( [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] created_port_ids = self._update_ports_for_instance( [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] with excutils.save_and_reraise_exception(): [ 619.086614] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.force_reraise() [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise self.value [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] updated_port = self._update_port( [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] _ensure_no_port_binding_failure(port) [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise exception.PortBindingFailed(port_id=port['id']) [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 619.086906] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] [ 619.086906] env[61907]: INFO nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Terminating instance [ 619.190449] env[61907]: DEBUG nova.network.neutron [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.274873] env[61907]: DEBUG nova.network.neutron [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.394492] env[61907]: DEBUG nova.network.neutron [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.397390] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 619.425184] env[61907]: DEBUG nova.scheduler.client.report [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 619.439224] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 619.439224] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 619.439666] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 619.439666] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 619.439666] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 619.439829] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 619.439961] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 619.440215] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 619.440403] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 619.440684] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 619.440757] env[61907]: DEBUG nova.virt.hardware [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.443336] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec37ae8-b246-4aab-bd87-a058993c03f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.450888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1511aa-9cdd-4c89-9b3b-99dc62a94726 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.595772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquiring lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.696709] env[61907]: DEBUG oslo_concurrency.lockutils [req-1f540618-a310-4f26-9362-d0ece3ad4f5c req-47687c13-73dd-4cd4-8c18-a89cb6064bb5 service nova] Releasing lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.697166] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquired lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.697354] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.901826] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Releasing lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.902294] env[61907]: DEBUG nova.compute.manager [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 619.902498] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.903395] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b596047b-6131-475f-b51f-82ad0281ca52 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.911181] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 619.911424] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2b2d129-a78f-40d0-9659-63d293c86933 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.919342] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 619.919342] env[61907]: value = "task-1243602" [ 619.919342] env[61907]: _type = "Task" [ 619.919342] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.931230] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.932377] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 619.937605] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243602, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.939845] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.519s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.941939] env[61907]: INFO nova.compute.claims [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 620.130010] env[61907]: DEBUG nova.compute.manager [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Received event network-changed-20e74284-c6f7-4668-bdbd-161482afcbcb {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 620.130277] env[61907]: DEBUG nova.compute.manager [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Refreshing instance network info cache due to event network-changed-20e74284-c6f7-4668-bdbd-161482afcbcb. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 620.130786] env[61907]: DEBUG oslo_concurrency.lockutils [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] Acquiring lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.130786] env[61907]: DEBUG oslo_concurrency.lockutils [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] Acquired lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.130786] env[61907]: DEBUG nova.network.neutron [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Refreshing network info cache for port 20e74284-c6f7-4668-bdbd-161482afcbcb {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.220796] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.352565] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.427033] env[61907]: DEBUG nova.compute.manager [req-c88983b6-b3a0-424b-8d46-886b75104ee8 req-969fee75-c87c-4c39-8ef9-0761dcff009e service nova] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Received event network-vif-deleted-a4de15f4-98c4-4297-8476-461ece4ed7a2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 620.433816] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243602, 'name': PowerOffVM_Task, 'duration_secs': 0.11333} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.434153] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 620.434796] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 620.434796] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6af71ca4-7371-4cfd-a72f-9f2243981594 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.440169] env[61907]: DEBUG nova.compute.utils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 620.445534] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 620.445534] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 620.458044] env[61907]: ERROR nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 620.458044] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.458044] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.458044] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.458044] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.458044] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.458044] env[61907]: ERROR nova.compute.manager raise self.value [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.458044] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 620.458044] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.458044] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 620.459138] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.459138] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 620.459138] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 620.459138] env[61907]: ERROR nova.compute.manager [ 620.459138] env[61907]: Traceback (most recent call last): [ 620.459138] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 620.459138] env[61907]: listener.cb(fileno) [ 620.459138] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.459138] env[61907]: result = function(*args, **kwargs) [ 620.459138] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.459138] env[61907]: return func(*args, **kwargs) [ 620.459138] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.459138] env[61907]: raise e [ 620.459138] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.459138] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 620.459138] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.459138] env[61907]: created_port_ids = self._update_ports_for_instance( [ 620.459138] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.459138] env[61907]: with excutils.save_and_reraise_exception(): [ 620.459138] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.459138] env[61907]: self.force_reraise() [ 620.459138] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.459138] env[61907]: raise self.value [ 620.459138] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.459138] env[61907]: updated_port = self._update_port( [ 620.459138] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.459138] env[61907]: _ensure_no_port_binding_failure(port) [ 620.459138] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.459138] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 620.459747] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 620.459747] env[61907]: Removing descriptor: 15 [ 620.460860] env[61907]: ERROR nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Traceback (most recent call last): [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] yield resources [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.driver.spawn(context, instance, image_meta, [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] vm_ref = self.build_virtual_machine(instance, [ 620.460860] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] for vif in network_info: [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self._sync_wrapper(fn, *args, **kwargs) [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.wait() [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self[:] = self._gt.wait() [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self._exit_event.wait() [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 620.461244] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] result = hub.switch() [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self.greenlet.switch() [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] result = function(*args, **kwargs) [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return func(*args, **kwargs) [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise e [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] nwinfo = self.network_api.allocate_for_instance( [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] created_port_ids = self._update_ports_for_instance( [ 620.461517] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] with excutils.save_and_reraise_exception(): [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.force_reraise() [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise self.value [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] updated_port = self._update_port( [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] _ensure_no_port_binding_failure(port) [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise exception.PortBindingFailed(port_id=port['id']) [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 620.461857] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] [ 620.462202] env[61907]: INFO nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Terminating instance [ 620.464762] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 620.464987] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 620.465191] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleting the datastore file [datastore2] 48b683f0-9cfe-472c-97df-1c19c3b5e441 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 620.468213] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cfa32042-6d7d-4ff1-9d9e-b82dfb214ecf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.481312] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for the task: (returnval){ [ 620.481312] env[61907]: value = "task-1243604" [ 620.481312] env[61907]: _type = "Task" [ 620.481312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.493209] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243604, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.518021] env[61907]: DEBUG nova.policy [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2b8c34677ab4906935c24795c6a652f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d09de37d6ff4c24a6e1a6367a236f08', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 620.656276] env[61907]: DEBUG nova.network.neutron [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.858201] env[61907]: DEBUG nova.network.neutron [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.860563] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Releasing lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.860563] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 620.861068] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-444a4d75-4425-4de1-8257-4df7b1510ec8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.871640] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10dd00b-d723-4fae-9b73-ab8cfb62ca38 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.896479] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 60d87612-650c-48dd-8afb-e0dc181c9c70 could not be found. [ 620.896733] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.897032] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-816cf25c-c71f-4cab-8759-1deb856b79d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.904550] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745fad0a-c421-4de1-8f4d-8adc5c5e88c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.927375] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 60d87612-650c-48dd-8afb-e0dc181c9c70 could not be found. [ 620.928238] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.928238] env[61907]: INFO nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Took 0.07 seconds to destroy the instance on the hypervisor. [ 620.928238] env[61907]: DEBUG oslo.service.loopingcall [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.928470] env[61907]: DEBUG nova.compute.manager [-] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 620.928470] env[61907]: DEBUG nova.network.neutron [-] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.945501] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 620.956834] env[61907]: DEBUG nova.network.neutron [-] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.975557] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 620.992896] env[61907]: DEBUG oslo_vmware.api [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Task: {'id': task-1243604, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112853} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.992896] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 620.992896] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 620.992896] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.993215] env[61907]: INFO nova.compute.manager [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Took 1.09 seconds to destroy the instance on the hypervisor. [ 620.993270] env[61907]: DEBUG oslo.service.loopingcall [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.993408] env[61907]: DEBUG nova.compute.manager [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 620.993497] env[61907]: DEBUG nova.network.neutron [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 621.016713] env[61907]: DEBUG nova.network.neutron [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.024297] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Successfully created port: 384066c6-dfea-4a6e-bd95-81ae85b9e253 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.362383] env[61907]: DEBUG oslo_concurrency.lockutils [req-69bedcbf-cd04-4d8a-ac81-fa81466bfaca req-eaf80db2-aca9-4a7b-9e4e-554ffc5c0b3d service nova] Releasing lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.362818] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquired lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.363017] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.404425] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.404674] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.444466] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc2f51c-5b4a-4288-96ab-d4548b7a2261 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.452161] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d61c96-62b7-4dc9-a033-92840712573b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.458751] env[61907]: DEBUG nova.network.neutron [-] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.496192] env[61907]: INFO nova.compute.manager [-] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Took 0.57 seconds to deallocate network for instance. [ 621.496868] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec77101-2042-4e24-a2fb-26497b08e637 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.510215] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956a137d-7669-46e7-9d69-5c2b48716701 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.527582] env[61907]: DEBUG nova.network.neutron [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.529517] env[61907]: DEBUG nova.compute.provider_tree [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.891737] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.963053] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 622.003693] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.003693] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.003693] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.004041] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.004041] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.004041] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.004041] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.004041] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.004165] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.004165] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.004165] env[61907]: DEBUG nova.virt.hardware [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.004165] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68854a8f-3619-44e4-bce5-b4acf7dbcd44 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.011014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55aa116e-7b92-4361-9ce1-abe66dddbc24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.026434] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.026434] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.031774] env[61907]: INFO nova.compute.manager [-] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Took 1.04 seconds to deallocate network for instance. [ 622.041611] env[61907]: DEBUG nova.scheduler.client.report [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 622.072208] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.072208] env[61907]: INFO nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Took 0.57 seconds to detach 1 volumes for instance. [ 622.076763] env[61907]: DEBUG nova.compute.claims [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 622.076954] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.237712] env[61907]: DEBUG nova.compute.manager [req-12cc23ee-0a4a-44a2-af64-9c7d11adbdd6 req-a937a7f6-5d60-4038-bfd0-63dff5c8fb8a service nova] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Received event network-vif-deleted-20e74284-c6f7-4668-bdbd-161482afcbcb {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 622.536252] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 622.537071] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 622.537071] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 622.547481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.609s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.548270] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 622.557135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.557135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.089s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.576115] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Releasing lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.577126] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 622.577126] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.577126] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c0503f6-dc59-4f86-8e99-c1dfcc66c8ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.587427] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3cd66d-ef55-45ec-a14e-dd2cc1645e7b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.612282] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92fa087f-a147-438b-9c30-c93baa5828dc could not be found. [ 622.612483] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.612683] env[61907]: INFO nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Took 0.04 seconds to destroy the instance on the hypervisor. [ 622.612960] env[61907]: DEBUG oslo.service.loopingcall [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.614162] env[61907]: DEBUG nova.compute.manager [-] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 622.614262] env[61907]: DEBUG nova.network.neutron [-] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.647354] env[61907]: DEBUG nova.network.neutron [-] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.924427] env[61907]: ERROR nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 622.924427] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.924427] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.924427] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.924427] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.924427] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.924427] env[61907]: ERROR nova.compute.manager raise self.value [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.924427] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.924427] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.924427] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.924893] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.924893] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.924893] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 622.924893] env[61907]: ERROR nova.compute.manager [ 622.924893] env[61907]: Traceback (most recent call last): [ 622.924893] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.924893] env[61907]: listener.cb(fileno) [ 622.924893] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.924893] env[61907]: result = function(*args, **kwargs) [ 622.924893] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.924893] env[61907]: return func(*args, **kwargs) [ 622.924893] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 622.924893] env[61907]: raise e [ 622.924893] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.924893] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 622.924893] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.924893] env[61907]: created_port_ids = self._update_ports_for_instance( [ 622.924893] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.924893] env[61907]: with excutils.save_and_reraise_exception(): [ 622.924893] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.924893] env[61907]: self.force_reraise() [ 622.924893] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.924893] env[61907]: raise self.value [ 622.924893] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.924893] env[61907]: updated_port = self._update_port( [ 622.924893] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.924893] env[61907]: _ensure_no_port_binding_failure(port) [ 622.924893] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.924893] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.925639] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 622.925639] env[61907]: Removing descriptor: 17 [ 622.925639] env[61907]: ERROR nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Traceback (most recent call last): [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] yield resources [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.driver.spawn(context, instance, image_meta, [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.925639] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] vm_ref = self.build_virtual_machine(instance, [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] for vif in network_info: [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self._sync_wrapper(fn, *args, **kwargs) [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.wait() [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self[:] = self._gt.wait() [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self._exit_event.wait() [ 622.925957] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] result = hub.switch() [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self.greenlet.switch() [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] result = function(*args, **kwargs) [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return func(*args, **kwargs) [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise e [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] nwinfo = self.network_api.allocate_for_instance( [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 622.926340] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] created_port_ids = self._update_ports_for_instance( [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] with excutils.save_and_reraise_exception(): [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.force_reraise() [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise self.value [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] updated_port = self._update_port( [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] _ensure_no_port_binding_failure(port) [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.926666] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise exception.PortBindingFailed(port_id=port['id']) [ 622.927016] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 622.927016] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] [ 622.927016] env[61907]: INFO nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Terminating instance [ 623.043627] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.044373] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.044583] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.044878] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.045066] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.045286] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.045428] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.045618] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 623.056984] env[61907]: DEBUG nova.compute.utils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 623.058152] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 623.058284] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 623.098373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.098373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.098373] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 623.098373] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 48b683f0-9cfe-472c-97df-1c19c3b5e441 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 623.150020] env[61907]: DEBUG nova.network.neutron [-] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.170316] env[61907]: DEBUG nova.policy [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4173a5d047834c1bacb2bf739e1973de', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f989a798a8524a2b99f7d9c5b7aeea15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 623.435977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquiring lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.436204] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquired lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.436406] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.564889] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 623.578161] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Successfully created port: 6a8682b1-c166-4fab-bc29-e0271db91e3a {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 623.586440] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1dd4d4-c273-44a1-a023-704c656b267f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.595427] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200fdf51-6804-4e6e-ad4e-b6be84447630 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.632975] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66958295-9df7-496e-9793-d4bd4ebebb0e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.642520] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84bf355-a061-460a-9a44-6c84e4c65c99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.659795] env[61907]: INFO nova.compute.manager [-] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Took 1.05 seconds to deallocate network for instance. [ 623.660423] env[61907]: DEBUG nova.compute.provider_tree [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.664953] env[61907]: DEBUG nova.compute.claims [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.664953] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.595877] env[61907]: DEBUG nova.scheduler.client.report [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 624.597955] env[61907]: DEBUG nova.compute.manager [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Received event network-changed-384066c6-dfea-4a6e-bd95-81ae85b9e253 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 624.597955] env[61907]: DEBUG nova.compute.manager [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Refreshing instance network info cache due to event network-changed-384066c6-dfea-4a6e-bd95-81ae85b9e253. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 624.599164] env[61907]: DEBUG oslo_concurrency.lockutils [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] Acquiring lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.626754] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.718219] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.865098] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.009582] env[61907]: ERROR nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 625.009582] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.009582] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.009582] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.009582] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.009582] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.009582] env[61907]: ERROR nova.compute.manager raise self.value [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.009582] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 625.009582] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.009582] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 625.010235] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.010235] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 625.010235] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 625.010235] env[61907]: ERROR nova.compute.manager [ 625.010235] env[61907]: Traceback (most recent call last): [ 625.010235] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 625.010235] env[61907]: listener.cb(fileno) [ 625.010235] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.010235] env[61907]: result = function(*args, **kwargs) [ 625.010235] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.010235] env[61907]: return func(*args, **kwargs) [ 625.010235] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.010235] env[61907]: raise e [ 625.010235] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.010235] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 625.010235] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.010235] env[61907]: created_port_ids = self._update_ports_for_instance( [ 625.010235] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.010235] env[61907]: with excutils.save_and_reraise_exception(): [ 625.010235] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.010235] env[61907]: self.force_reraise() [ 625.010235] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.010235] env[61907]: raise self.value [ 625.010235] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.010235] env[61907]: updated_port = self._update_port( [ 625.010235] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.010235] env[61907]: _ensure_no_port_binding_failure(port) [ 625.010235] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.010235] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 625.010923] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 625.010923] env[61907]: Removing descriptor: 15 [ 625.100439] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 625.103196] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.548s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.106833] env[61907]: ERROR nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Traceback (most recent call last): [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.driver.spawn(context, instance, image_meta, [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] vm_ref = self.build_virtual_machine(instance, [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.106833] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] for vif in network_info: [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self._sync_wrapper(fn, *args, **kwargs) [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.wait() [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self[:] = self._gt.wait() [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self._exit_event.wait() [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] result = hub.switch() [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 625.107775] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return self.greenlet.switch() [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] result = function(*args, **kwargs) [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] return func(*args, **kwargs) [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise e [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] nwinfo = self.network_api.allocate_for_instance( [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] created_port_ids = self._update_ports_for_instance( [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] with excutils.save_and_reraise_exception(): [ 625.108098] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] self.force_reraise() [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise self.value [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] updated_port = self._update_port( [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] _ensure_no_port_binding_failure(port) [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] raise exception.PortBindingFailed(port_id=port['id']) [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] nova.exception.PortBindingFailed: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. [ 625.108406] env[61907]: ERROR nova.compute.manager [instance: bed89649-b8d3-414a-8729-a504a7e02b94] [ 625.108734] env[61907]: DEBUG nova.compute.utils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 625.113404] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.337s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.116035] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Build of instance bed89649-b8d3-414a-8729-a504a7e02b94 was re-scheduled: Binding failed for port 7a62a6ed-b052-4f40-ae09-f63fec67cd44, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 625.118271] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 625.118271] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquiring lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.118271] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Acquired lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.118271] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.148145] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 625.149031] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 625.149031] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 625.149031] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 625.149031] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 625.152556] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 625.156025] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 625.156025] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 625.156025] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 625.156025] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 625.156025] env[61907]: DEBUG nova.virt.hardware [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 625.156227] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409df6f1-e848-4393-b667-ef2dabc96990 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.167248] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0acef216-501b-4590-8161-322059031d9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.189681] env[61907]: ERROR nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] Traceback (most recent call last): [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] yield resources [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.driver.spawn(context, instance, image_meta, [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] vm_ref = self.build_virtual_machine(instance, [ 625.189681] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] for vif in network_info: [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return self._sync_wrapper(fn, *args, **kwargs) [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.wait() [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self[:] = self._gt.wait() [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return self._exit_event.wait() [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 625.190031] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] current.throw(*self._exc) [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] result = function(*args, **kwargs) [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return func(*args, **kwargs) [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise e [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] nwinfo = self.network_api.allocate_for_instance( [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] created_port_ids = self._update_ports_for_instance( [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] with excutils.save_and_reraise_exception(): [ 625.190355] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.force_reraise() [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise self.value [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] updated_port = self._update_port( [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] _ensure_no_port_binding_failure(port) [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise exception.PortBindingFailed(port_id=port['id']) [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 625.190822] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] [ 625.190822] env[61907]: INFO nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Terminating instance [ 625.324642] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.372739] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Releasing lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.373099] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 625.373273] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 625.373628] env[61907]: DEBUG oslo_concurrency.lockutils [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] Acquired lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.373832] env[61907]: DEBUG nova.network.neutron [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Refreshing network info cache for port 384066c6-dfea-4a6e-bd95-81ae85b9e253 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.375419] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97d21425-2f43-4ed2-ae30-fe7799931cbb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.392409] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff453de-cf8c-4bf0-ae59-81af4b1e8301 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.425233] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 420a661e-33aa-43e5-b298-19a20e8ab33b could not be found. [ 625.425233] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 625.425233] env[61907]: INFO nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Took 0.05 seconds to destroy the instance on the hypervisor. [ 625.425233] env[61907]: DEBUG oslo.service.loopingcall [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 625.425233] env[61907]: DEBUG nova.compute.manager [-] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 625.425233] env[61907]: DEBUG nova.network.neutron [-] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 625.465302] env[61907]: DEBUG nova.network.neutron [-] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.653215] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.697023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquiring lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.697023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquired lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.697023] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 625.826967] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-48b683f0-9cfe-472c-97df-1c19c3b5e441" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.827209] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 625.827410] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.827571] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.827755] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.827915] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.828067] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.828216] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.828366] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 625.828523] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 625.923873] env[61907]: DEBUG nova.network.neutron [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.973624] env[61907]: DEBUG nova.network.neutron [-] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.041388] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.048506] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f685f2b-1b60-49d4-b35a-aeb6561c0b57 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.059789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b67ec62-984c-4467-bddb-b4fa1dd821ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.092013] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37a80cb6-78e6-4f5a-a889-e3b45690cf78 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.099717] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44855540-7466-4d4b-a926-2a7be124898b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.114446] env[61907]: DEBUG nova.compute.provider_tree [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.143852] env[61907]: DEBUG nova.network.neutron [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.224939] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.332125] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.367058] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.478076] env[61907]: INFO nova.compute.manager [-] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Took 1.05 seconds to deallocate network for instance. [ 626.480830] env[61907]: DEBUG nova.compute.claims [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 626.481021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.500332] env[61907]: DEBUG nova.compute.manager [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Received event network-changed-6a8682b1-c166-4fab-bc29-e0271db91e3a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 626.500608] env[61907]: DEBUG nova.compute.manager [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Refreshing instance network info cache due to event network-changed-6a8682b1-c166-4fab-bc29-e0271db91e3a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 626.500736] env[61907]: DEBUG oslo_concurrency.lockutils [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] Acquiring lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.541520] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Releasing lock "refresh_cache-bed89649-b8d3-414a-8729-a504a7e02b94" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.541916] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 626.542066] env[61907]: DEBUG nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.542235] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.576022] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.617706] env[61907]: DEBUG nova.scheduler.client.report [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 626.646154] env[61907]: DEBUG oslo_concurrency.lockutils [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] Releasing lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.646573] env[61907]: DEBUG nova.compute.manager [req-164790ce-22e5-415b-af26-844c2af367d1 req-377c7772-52de-4e3b-b6d0-df00541eea79 service nova] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Received event network-vif-deleted-384066c6-dfea-4a6e-bd95-81ae85b9e253 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 626.870744] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Releasing lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.871117] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 626.871284] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 626.871619] env[61907]: DEBUG oslo_concurrency.lockutils [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] Acquired lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.871995] env[61907]: DEBUG nova.network.neutron [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Refreshing network info cache for port 6a8682b1-c166-4fab-bc29-e0271db91e3a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.873512] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8bd61ca0-25eb-4315-afc3-3aff3589acd8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.883935] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd8f391c-4e36-4e24-b7f8-83138447dec4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.910746] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a7e883a1-aab9-4372-8379-144b373062e2 could not be found. [ 626.910918] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 626.911119] env[61907]: INFO nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 626.911362] env[61907]: DEBUG oslo.service.loopingcall [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 626.911586] env[61907]: DEBUG nova.compute.manager [-] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 626.911760] env[61907]: DEBUG nova.network.neutron [-] [instance: a7e883a1-aab9-4372-8379-144b373062e2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 626.931964] env[61907]: DEBUG nova.network.neutron [-] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.940305] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquiring lock "f517b2e4-1a83-4519-8a5b-65d489d16b6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.940676] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "f517b2e4-1a83-4519-8a5b-65d489d16b6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.078985] env[61907]: DEBUG nova.network.neutron [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.127577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.128223] env[61907]: ERROR nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Traceback (most recent call last): [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.driver.spawn(context, instance, image_meta, [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] vm_ref = self.build_virtual_machine(instance, [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.128223] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] for vif in network_info: [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self._sync_wrapper(fn, *args, **kwargs) [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.wait() [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self[:] = self._gt.wait() [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self._exit_event.wait() [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] result = hub.switch() [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.128976] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return self.greenlet.switch() [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] result = function(*args, **kwargs) [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] return func(*args, **kwargs) [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise e [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] nwinfo = self.network_api.allocate_for_instance( [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] created_port_ids = self._update_ports_for_instance( [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] with excutils.save_and_reraise_exception(): [ 627.129515] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] self.force_reraise() [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise self.value [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] updated_port = self._update_port( [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] _ensure_no_port_binding_failure(port) [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] raise exception.PortBindingFailed(port_id=port['id']) [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] nova.exception.PortBindingFailed: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. [ 627.130119] env[61907]: ERROR nova.compute.manager [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] [ 627.131464] env[61907]: DEBUG nova.compute.utils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.131464] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Build of instance 5556aefa-3c0b-415c-94f3-603c7fb26ec1 was re-scheduled: Binding failed for port a2f642bb-14c2-4767-8857-41f4dd1fe745, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 627.131464] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 627.131464] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.131607] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquired lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.131673] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.134762] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.851s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.134948] env[61907]: INFO nova.compute.claims [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 627.395270] env[61907]: DEBUG nova.network.neutron [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.436380] env[61907]: DEBUG nova.network.neutron [-] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.531572] env[61907]: DEBUG nova.network.neutron [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.582622] env[61907]: INFO nova.compute.manager [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] [instance: bed89649-b8d3-414a-8729-a504a7e02b94] Took 1.04 seconds to deallocate network for instance. [ 627.653963] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.742349] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.940017] env[61907]: INFO nova.compute.manager [-] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Took 1.03 seconds to deallocate network for instance. [ 627.942544] env[61907]: DEBUG nova.compute.claims [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 627.942720] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.034295] env[61907]: DEBUG oslo_concurrency.lockutils [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] Releasing lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.034496] env[61907]: DEBUG nova.compute.manager [req-368c78f1-a2d4-4f00-8f6d-90b8ee80ed58 req-006f820c-32e6-4d69-b2de-4e0812fc2115 service nova] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Received event network-vif-deleted-6a8682b1-c166-4fab-bc29-e0271db91e3a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 628.245285] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Releasing lock "refresh_cache-5556aefa-3c0b-415c-94f3-603c7fb26ec1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.245523] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 628.245700] env[61907]: DEBUG nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 628.245865] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.264526] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.541975] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-914f255f-0f4c-47b1-835d-2ba15e2533cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.549408] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6645b5-de67-45d1-953b-cf326ff70521 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.578770] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97232955-bc5b-4d0c-95af-ecb88bf37cdd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.586221] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719f2285-71e9-4335-87b5-015513829471 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.602433] env[61907]: DEBUG nova.compute.provider_tree [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.624367] env[61907]: INFO nova.scheduler.client.report [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Deleted allocations for instance bed89649-b8d3-414a-8729-a504a7e02b94 [ 628.767362] env[61907]: DEBUG nova.network.neutron [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.105875] env[61907]: DEBUG nova.scheduler.client.report [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 629.132756] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7a5a0aa6-5926-4bab-9d47-4daaef2ed568 tempest-AttachInterfacesV270Test-1676923903 tempest-AttachInterfacesV270Test-1676923903-project-member] Lock "bed89649-b8d3-414a-8729-a504a7e02b94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.580s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.271020] env[61907]: INFO nova.compute.manager [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 5556aefa-3c0b-415c-94f3-603c7fb26ec1] Took 1.02 seconds to deallocate network for instance. [ 629.611707] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.611707] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 629.613952] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.150s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.635194] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 630.127307] env[61907]: DEBUG nova.compute.utils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 630.128780] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 630.129150] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 630.165119] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.199720] env[61907]: DEBUG nova.policy [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 630.321722] env[61907]: INFO nova.scheduler.client.report [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Deleted allocations for instance 5556aefa-3c0b-415c-94f3-603c7fb26ec1 [ 630.633749] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 630.637669] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Successfully created port: 2db53abe-3598-498f-b4d7-2339c61124a6 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.644023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b765e3d-80ae-4b69-a001-07e694d8372a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.659771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34aabd2-fdad-4e2a-8c75-e1a0f1ca38d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.693882] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23dea14-40c2-479c-b657-69d5c8e0ac0b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.701922] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74ae1b71-a574-4cdd-a578-3b31ae8b105a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.716520] env[61907]: DEBUG nova.compute.provider_tree [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.832681] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89e2a231-9ae4-4a1e-9b3b-58fd09112369 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "5556aefa-3c0b-415c-94f3-603c7fb26ec1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.104s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.219213] env[61907]: DEBUG nova.scheduler.client.report [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 631.337362] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 631.489402] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquiring lock "1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.489740] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.650126] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 631.686120] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.686368] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.686529] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.686711] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.686869] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.687068] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.687287] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.687445] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.688669] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.688669] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.688669] env[61907]: DEBUG nova.virt.hardware [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.688916] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3c7f7b-a7cd-48ea-8627-b70863fdf7f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.696971] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8314e3c-7214-400a-9b0e-c2882503db10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.724325] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.110s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.725026] env[61907]: ERROR nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Traceback (most recent call last): [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.driver.spawn(context, instance, image_meta, [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] vm_ref = self.build_virtual_machine(instance, [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.725026] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] for vif in network_info: [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self._sync_wrapper(fn, *args, **kwargs) [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.wait() [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self[:] = self._gt.wait() [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self._exit_event.wait() [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] result = hub.switch() [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.725360] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return self.greenlet.switch() [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] result = function(*args, **kwargs) [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] return func(*args, **kwargs) [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise e [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] nwinfo = self.network_api.allocate_for_instance( [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] created_port_ids = self._update_ports_for_instance( [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] with excutils.save_and_reraise_exception(): [ 631.725720] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] self.force_reraise() [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise self.value [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] updated_port = self._update_port( [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] _ensure_no_port_binding_failure(port) [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] raise exception.PortBindingFailed(port_id=port['id']) [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] nova.exception.PortBindingFailed: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. [ 631.726040] env[61907]: ERROR nova.compute.manager [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] [ 631.726502] env[61907]: DEBUG nova.compute.utils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.727966] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.169s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.729634] env[61907]: INFO nova.compute.claims [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.733015] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Build of instance c6fd5d45-1abd-4a12-8fcf-86930161bafb was re-scheduled: Binding failed for port 8885367d-c594-44d9-9ae6-13c3f30bee98, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 631.734295] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 631.737873] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.737873] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.737873] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 631.870154] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.951717] env[61907]: DEBUG nova.compute.manager [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Received event network-changed-2db53abe-3598-498f-b4d7-2339c61124a6 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 631.951920] env[61907]: DEBUG nova.compute.manager [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Refreshing instance network info cache due to event network-changed-2db53abe-3598-498f-b4d7-2339c61124a6. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 631.952148] env[61907]: DEBUG oslo_concurrency.lockutils [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] Acquiring lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.952296] env[61907]: DEBUG oslo_concurrency.lockutils [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] Acquired lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.952437] env[61907]: DEBUG nova.network.neutron [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Refreshing network info cache for port 2db53abe-3598-498f-b4d7-2339c61124a6 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 632.114959] env[61907]: ERROR nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 632.114959] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.114959] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.114959] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.114959] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.114959] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.114959] env[61907]: ERROR nova.compute.manager raise self.value [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.114959] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.114959] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.114959] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.115409] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.115409] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.115409] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 632.115409] env[61907]: ERROR nova.compute.manager [ 632.115409] env[61907]: Traceback (most recent call last): [ 632.115409] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.115409] env[61907]: listener.cb(fileno) [ 632.115409] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.115409] env[61907]: result = function(*args, **kwargs) [ 632.115409] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.115409] env[61907]: return func(*args, **kwargs) [ 632.115409] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.115409] env[61907]: raise e [ 632.115409] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.115409] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 632.115409] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.115409] env[61907]: created_port_ids = self._update_ports_for_instance( [ 632.115409] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.115409] env[61907]: with excutils.save_and_reraise_exception(): [ 632.115409] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.115409] env[61907]: self.force_reraise() [ 632.115409] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.115409] env[61907]: raise self.value [ 632.115409] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.115409] env[61907]: updated_port = self._update_port( [ 632.115409] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.115409] env[61907]: _ensure_no_port_binding_failure(port) [ 632.115409] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.115409] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.116103] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 632.116103] env[61907]: Removing descriptor: 17 [ 632.116160] env[61907]: ERROR nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Traceback (most recent call last): [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] yield resources [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.driver.spawn(context, instance, image_meta, [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] vm_ref = self.build_virtual_machine(instance, [ 632.116160] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] for vif in network_info: [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self._sync_wrapper(fn, *args, **kwargs) [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.wait() [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self[:] = self._gt.wait() [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self._exit_event.wait() [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.116443] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] result = hub.switch() [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self.greenlet.switch() [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] result = function(*args, **kwargs) [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return func(*args, **kwargs) [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise e [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] nwinfo = self.network_api.allocate_for_instance( [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] created_port_ids = self._update_ports_for_instance( [ 632.116777] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] with excutils.save_and_reraise_exception(): [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.force_reraise() [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise self.value [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] updated_port = self._update_port( [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] _ensure_no_port_binding_failure(port) [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise exception.PortBindingFailed(port_id=port['id']) [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 632.117153] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] [ 632.117502] env[61907]: INFO nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Terminating instance [ 632.280265] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.428971] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.475763] env[61907]: DEBUG nova.network.neutron [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.604076] env[61907]: DEBUG nova.network.neutron [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.621531] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.934824] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "refresh_cache-c6fd5d45-1abd-4a12-8fcf-86930161bafb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.939121] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 632.939121] env[61907]: DEBUG nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 632.939121] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 632.965190] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.110657] env[61907]: DEBUG oslo_concurrency.lockutils [req-60e2954b-39bc-4abc-91fa-ce4d0caf4fe0 req-ae8c922a-046c-4f6c-a2d2-e9ef6a9781b9 service nova] Releasing lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.111763] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.112380] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.255019] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af940c3-f1c3-4e28-bb60-d1817e6273aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.262165] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3323d6a-ef0e-4c47-83eb-6d8e1a73513a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.298037] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de94c15-7d54-4189-b5ee-9f244558e787 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.305466] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e5f6909-e7c4-4cd7-b96a-ef0f5d44ea72 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.318721] env[61907]: DEBUG nova.compute.provider_tree [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.471351] env[61907]: DEBUG nova.network.neutron [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.636757] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.809969] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.823068] env[61907]: DEBUG nova.scheduler.client.report [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 633.976747] env[61907]: INFO nova.compute.manager [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: c6fd5d45-1abd-4a12-8fcf-86930161bafb] Took 1.04 seconds to deallocate network for instance. [ 634.085338] env[61907]: DEBUG nova.compute.manager [req-63ebd01f-87b8-4228-ac81-c1cc9189b9d3 req-2041267f-c13b-460a-b96a-af39945d4094 service nova] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Received event network-vif-deleted-2db53abe-3598-498f-b4d7-2339c61124a6 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 634.312988] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.313524] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 634.313792] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.314121] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d8426dd-e182-458e-a324-e9f8f54e0a67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.323361] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2cddd52-cb5e-4975-8c85-ea44c7a15fc3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.334809] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.607s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.335341] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 634.337825] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.724s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.354660] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d95c4916-660e-4ead-ba87-2f69871d5f1e could not be found. [ 634.354901] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.355095] env[61907]: INFO nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.355337] env[61907]: DEBUG oslo.service.loopingcall [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.355545] env[61907]: DEBUG nova.compute.manager [-] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 634.355639] env[61907]: DEBUG nova.network.neutron [-] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.374122] env[61907]: DEBUG nova.network.neutron [-] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.652213] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "215d0f35-5d19-4db4-996d-571e2970e19c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.652455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "215d0f35-5d19-4db4-996d-571e2970e19c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.844118] env[61907]: DEBUG nova.compute.utils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.849018] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 634.849018] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 634.879431] env[61907]: DEBUG nova.network.neutron [-] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.905158] env[61907]: DEBUG nova.policy [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '12e6e3d6b4d54da1b39b1b9ba210d4af', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b1e1197870d94715bf949e49ca6ce9f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.010746] env[61907]: INFO nova.scheduler.client.report [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted allocations for instance c6fd5d45-1abd-4a12-8fcf-86930161bafb [ 635.348485] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 635.373958] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Successfully created port: d5f5e136-ba67-41a9-bce1-23cff18cb55f {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 635.382763] env[61907]: INFO nova.compute.manager [-] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Took 1.03 seconds to deallocate network for instance. [ 635.388015] env[61907]: DEBUG nova.compute.claims [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.388015] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.412741] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18846983-1308-43e2-b0a5-7f9c14384a8f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.421124] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2147208f-2eb2-4454-934a-56a318a25b6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.458517] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccbdb27-1a2a-4665-a611-1f30a18db560 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.466330] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf23b031-ceeb-45c4-93a2-9c4d174bc5df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.481279] env[61907]: DEBUG nova.compute.provider_tree [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.525889] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5db0db00-9268-45a3-9144-ecd67cd42815 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "c6fd5d45-1abd-4a12-8fcf-86930161bafb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.576s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.984228] env[61907]: DEBUG nova.scheduler.client.report [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 636.029399] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 636.361942] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 636.388409] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.388869] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.389080] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.389392] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.389772] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.389973] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.390259] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.390614] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.393330] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.393330] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.393330] env[61907]: DEBUG nova.virt.hardware [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.393330] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79805e0f-798c-4176-a9dd-d2d8309c4d17 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.404016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33180319-74c6-4286-a798-43901a916f06 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.488977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.151s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.489693] env[61907]: ERROR nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Traceback (most recent call last): [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.driver.spawn(context, instance, image_meta, [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] vm_ref = self.build_virtual_machine(instance, [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.489693] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] for vif in network_info: [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self._sync_wrapper(fn, *args, **kwargs) [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.wait() [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self[:] = self._gt.wait() [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self._exit_event.wait() [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] result = hub.switch() [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.490642] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return self.greenlet.switch() [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] result = function(*args, **kwargs) [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] return func(*args, **kwargs) [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise e [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] nwinfo = self.network_api.allocate_for_instance( [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] created_port_ids = self._update_ports_for_instance( [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] with excutils.save_and_reraise_exception(): [ 636.491113] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] self.force_reraise() [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise self.value [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] updated_port = self._update_port( [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] _ensure_no_port_binding_failure(port) [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] raise exception.PortBindingFailed(port_id=port['id']) [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] nova.exception.PortBindingFailed: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. [ 636.491960] env[61907]: ERROR nova.compute.manager [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] [ 636.492523] env[61907]: DEBUG nova.compute.utils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.492523] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.415s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.496604] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Build of instance 2fcb4164-c0c5-49e5-87cb-a51df711b5a2 was re-scheduled: Binding failed for port 4349e1bc-b7ad-4ceb-8e38-276020ab0189, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 636.496604] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 636.496604] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquiring lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.496604] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Acquired lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.496753] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.554029] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.637462] env[61907]: DEBUG nova.compute.manager [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Received event network-changed-d5f5e136-ba67-41a9-bce1-23cff18cb55f {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 636.637575] env[61907]: DEBUG nova.compute.manager [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Refreshing instance network info cache due to event network-changed-d5f5e136-ba67-41a9-bce1-23cff18cb55f. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 636.637805] env[61907]: DEBUG oslo_concurrency.lockutils [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] Acquiring lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.637968] env[61907]: DEBUG oslo_concurrency.lockutils [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] Acquired lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.638164] env[61907]: DEBUG nova.network.neutron [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Refreshing network info cache for port d5f5e136-ba67-41a9-bce1-23cff18cb55f {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 636.728284] env[61907]: ERROR nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 636.728284] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.728284] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.728284] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.728284] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.728284] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.728284] env[61907]: ERROR nova.compute.manager raise self.value [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.728284] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 636.728284] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.728284] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 636.728816] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.728816] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 636.728816] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 636.728816] env[61907]: ERROR nova.compute.manager [ 636.728816] env[61907]: Traceback (most recent call last): [ 636.728816] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 636.728816] env[61907]: listener.cb(fileno) [ 636.728816] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.728816] env[61907]: result = function(*args, **kwargs) [ 636.728816] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.728816] env[61907]: return func(*args, **kwargs) [ 636.728816] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.728816] env[61907]: raise e [ 636.728816] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.728816] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 636.728816] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.728816] env[61907]: created_port_ids = self._update_ports_for_instance( [ 636.728816] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.728816] env[61907]: with excutils.save_and_reraise_exception(): [ 636.728816] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.728816] env[61907]: self.force_reraise() [ 636.728816] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.728816] env[61907]: raise self.value [ 636.728816] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.728816] env[61907]: updated_port = self._update_port( [ 636.728816] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.728816] env[61907]: _ensure_no_port_binding_failure(port) [ 636.728816] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.728816] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 636.729649] env[61907]: nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 636.729649] env[61907]: Removing descriptor: 17 [ 636.729649] env[61907]: ERROR nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Traceback (most recent call last): [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] yield resources [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.driver.spawn(context, instance, image_meta, [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.729649] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] vm_ref = self.build_virtual_machine(instance, [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] for vif in network_info: [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self._sync_wrapper(fn, *args, **kwargs) [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.wait() [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self[:] = self._gt.wait() [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self._exit_event.wait() [ 636.729976] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] result = hub.switch() [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self.greenlet.switch() [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] result = function(*args, **kwargs) [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return func(*args, **kwargs) [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise e [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] nwinfo = self.network_api.allocate_for_instance( [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.730341] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] created_port_ids = self._update_ports_for_instance( [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] with excutils.save_and_reraise_exception(): [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.force_reraise() [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise self.value [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] updated_port = self._update_port( [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] _ensure_no_port_binding_failure(port) [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.731236] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise exception.PortBindingFailed(port_id=port['id']) [ 636.731597] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 636.731597] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] [ 636.731597] env[61907]: INFO nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Terminating instance [ 637.019613] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.130296] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.158357] env[61907]: DEBUG nova.network.neutron [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.233327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.260197] env[61907]: DEBUG nova.network.neutron [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.268243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquiring lock "45c58b31-d0cc-4b36-b744-448a38e11266" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.269035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "45c58b31-d0cc-4b36-b744-448a38e11266" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.452924] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe63c83-5f13-4ae1-9d49-e01b63e12c7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.462273] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10cef26e-4adf-4ab9-8c07-9dc5c57333b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.500382] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225f3613-aa71-47ca-9b5f-f762bf0cc5e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.508185] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b436201-449b-4ccf-a363-c0732293c6ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.521960] env[61907]: DEBUG nova.compute.provider_tree [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.635184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Releasing lock "refresh_cache-2fcb4164-c0c5-49e5-87cb-a51df711b5a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.637468] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 637.637468] env[61907]: DEBUG nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 637.637468] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.657280] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.765949] env[61907]: DEBUG oslo_concurrency.lockutils [req-9260952a-0ab2-43fb-b689-35d4a3715dcd req-4a5ac3fb-e04c-48a0-83f3-cfa37367e5cc service nova] Releasing lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.766261] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquired lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.766421] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.028019] env[61907]: DEBUG nova.scheduler.client.report [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 638.163170] env[61907]: DEBUG nova.network.neutron [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.262321] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.262321] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.284205] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.363493] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.531680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.039s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.532019] env[61907]: ERROR nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Traceback (most recent call last): [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.driver.spawn(context, instance, image_meta, [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] vm_ref = self.build_virtual_machine(instance, [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.532019] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] for vif in network_info: [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return self._sync_wrapper(fn, *args, **kwargs) [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.wait() [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self[:] = self._gt.wait() [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return self._exit_event.wait() [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] current.throw(*self._exc) [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.532356] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] result = function(*args, **kwargs) [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] return func(*args, **kwargs) [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise e [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] nwinfo = self.network_api.allocate_for_instance( [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] created_port_ids = self._update_ports_for_instance( [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] with excutils.save_and_reraise_exception(): [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] self.force_reraise() [ 638.532704] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise self.value [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] updated_port = self._update_port( [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] _ensure_no_port_binding_failure(port) [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] raise exception.PortBindingFailed(port_id=port['id']) [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] nova.exception.PortBindingFailed: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. [ 638.533116] env[61907]: ERROR nova.compute.manager [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] [ 638.533116] env[61907]: DEBUG nova.compute.utils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.534062] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.980s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.534283] env[61907]: DEBUG nova.objects.instance [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lazy-loading 'resources' on Instance uuid 48b683f0-9cfe-472c-97df-1c19c3b5e441 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 638.535472] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Build of instance 60d87612-650c-48dd-8afb-e0dc181c9c70 was re-scheduled: Binding failed for port a4de15f4-98c4-4297-8476-461ece4ed7a2, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 638.535863] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 638.536097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquiring lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.536236] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Acquired lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.536411] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.656437] env[61907]: DEBUG nova.compute.manager [req-6489c9ba-be9e-4df8-9ffd-fbe4b6dfca07 req-0df36441-d52d-449a-a49c-3674cd9e02bc service nova] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Received event network-vif-deleted-d5f5e136-ba67-41a9-bce1-23cff18cb55f {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 638.666455] env[61907]: INFO nova.compute.manager [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] [instance: 2fcb4164-c0c5-49e5-87cb-a51df711b5a2] Took 1.03 seconds to deallocate network for instance. [ 638.866047] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Releasing lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.866511] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 638.866712] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 638.867035] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4243085c-2d97-4b2f-bae0-15ca3336a257 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.875781] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5aa58378-394c-427f-b59a-618647dc3b82 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.897443] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe could not be found. [ 638.897684] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 638.898659] env[61907]: INFO nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.898954] env[61907]: DEBUG oslo.service.loopingcall [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.899208] env[61907]: DEBUG nova.compute.manager [-] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 638.899306] env[61907]: DEBUG nova.network.neutron [-] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.917029] env[61907]: DEBUG nova.network.neutron [-] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.060240] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.139070] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.398972] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a486c595-7ca9-49a2-bcfd-83c31892d86b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.406094] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6961925c-8c3b-4c9a-87e9-ed81183f23ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.434910] env[61907]: DEBUG nova.network.neutron [-] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.436771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a7a8a4-5625-49fc-af47-0b095064dd50 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.444375] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2267ef24-c7ec-4751-ae0c-5c98fa51c011 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.458187] env[61907]: DEBUG nova.compute.provider_tree [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.641208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Releasing lock "refresh_cache-60d87612-650c-48dd-8afb-e0dc181c9c70" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.641488] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 639.641639] env[61907]: DEBUG nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 639.641807] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.656627] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.694116] env[61907]: INFO nova.scheduler.client.report [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Deleted allocations for instance 2fcb4164-c0c5-49e5-87cb-a51df711b5a2 [ 639.942015] env[61907]: INFO nova.compute.manager [-] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Took 1.04 seconds to deallocate network for instance. [ 639.943710] env[61907]: DEBUG nova.compute.claims [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.944038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.961204] env[61907]: DEBUG nova.scheduler.client.report [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 640.159568] env[61907]: DEBUG nova.network.neutron [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.201209] env[61907]: DEBUG oslo_concurrency.lockutils [None req-888cb3d7-46b0-41d2-8b61-e08dc1688dd2 tempest-ServerAddressesNegativeTestJSON-586959205 tempest-ServerAddressesNegativeTestJSON-586959205-project-member] Lock "2fcb4164-c0c5-49e5-87cb-a51df711b5a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.868s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.468907] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.469306] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.804s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.496013] env[61907]: INFO nova.scheduler.client.report [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Deleted allocations for instance 48b683f0-9cfe-472c-97df-1c19c3b5e441 [ 640.662730] env[61907]: INFO nova.compute.manager [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] [instance: 60d87612-650c-48dd-8afb-e0dc181c9c70] Took 1.02 seconds to deallocate network for instance. [ 640.704013] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 641.012121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-83445479-b59c-428b-a152-d6ac761e137d tempest-ServerDiagnosticsV248Test-842860052 tempest-ServerDiagnosticsV248Test-842860052-project-member] Lock "48b683f0-9cfe-472c-97df-1c19c3b5e441" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.785s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.225374] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.416642] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b97bd5-1143-4573-bfb3-42a17f6f4a66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.426018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684e99a2-d77e-4ffe-8773-49a903735f4f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.454649] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca09a7c6-eec5-4673-bedb-a90f2fa2be78 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.462408] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34bb7e1c-1666-42fe-a0fe-613aa4daffe5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.479025] env[61907]: DEBUG nova.compute.provider_tree [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.705789] env[61907]: INFO nova.scheduler.client.report [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Deleted allocations for instance 60d87612-650c-48dd-8afb-e0dc181c9c70 [ 641.981093] env[61907]: DEBUG nova.scheduler.client.report [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 642.213282] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dba5a4b1-9d24-4046-8001-0ff2c8d90b81 tempest-ServersTestBootFromVolume-1005710906 tempest-ServersTestBootFromVolume-1005710906-project-member] Lock "60d87612-650c-48dd-8afb-e0dc181c9c70" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.974s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.488601] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.488601] env[61907]: ERROR nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Traceback (most recent call last): [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.driver.spawn(context, instance, image_meta, [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.488601] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] vm_ref = self.build_virtual_machine(instance, [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] for vif in network_info: [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self._sync_wrapper(fn, *args, **kwargs) [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.wait() [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self[:] = self._gt.wait() [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self._exit_event.wait() [ 642.489071] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] result = hub.switch() [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return self.greenlet.switch() [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] result = function(*args, **kwargs) [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] return func(*args, **kwargs) [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise e [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] nwinfo = self.network_api.allocate_for_instance( [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.492454] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] created_port_ids = self._update_ports_for_instance( [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] with excutils.save_and_reraise_exception(): [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] self.force_reraise() [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise self.value [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] updated_port = self._update_port( [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] _ensure_no_port_binding_failure(port) [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.492820] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] raise exception.PortBindingFailed(port_id=port['id']) [ 642.493179] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] nova.exception.PortBindingFailed: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. [ 642.493179] env[61907]: ERROR nova.compute.manager [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] [ 642.493179] env[61907]: DEBUG nova.compute.utils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.497500] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Build of instance 92fa087f-a147-438b-9c30-c93baa5828dc was re-scheduled: Binding failed for port 20e74284-c6f7-4668-bdbd-161482afcbcb, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 642.497500] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 642.497500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.497500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquired lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.497659] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.497659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.164s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.497659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.497659] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 642.497659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.016s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.502136] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d12ba310-c50a-4dd5-bbeb-907fba6c86fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.515789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91cc1a7-08b0-42e8-9e7a-497724ef8dcc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.531524] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6273ab39-154f-449f-a985-b4d819313389 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.538529] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a3d72f-1da8-4ce5-819a-01f0e0437fce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.569790] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181391MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 642.569790] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.717967] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 643.031874] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.173458] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.242447] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.480470] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03bd4034-bdc2-4f7e-b24d-e9507035b8ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.489696] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27efdf52-367e-4014-968e-d61fe2b709b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.524657] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7eb5aa-2c8d-4a84-a627-46f39ffdc7a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.532268] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0b73fe0-dbc1-4623-b10d-e8735cb0e617 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.549609] env[61907]: DEBUG nova.compute.provider_tree [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.678651] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Releasing lock "refresh_cache-92fa087f-a147-438b-9c30-c93baa5828dc" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.678651] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 643.678651] env[61907]: DEBUG nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 643.678651] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.700020] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.058111] env[61907]: DEBUG nova.scheduler.client.report [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 644.199543] env[61907]: DEBUG nova.network.neutron [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.564834] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.068s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.566256] env[61907]: ERROR nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Traceback (most recent call last): [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.driver.spawn(context, instance, image_meta, [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] vm_ref = self.build_virtual_machine(instance, [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.566256] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] for vif in network_info: [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self._sync_wrapper(fn, *args, **kwargs) [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.wait() [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self[:] = self._gt.wait() [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self._exit_event.wait() [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] result = hub.switch() [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.566966] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return self.greenlet.switch() [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] result = function(*args, **kwargs) [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] return func(*args, **kwargs) [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise e [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] nwinfo = self.network_api.allocate_for_instance( [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] created_port_ids = self._update_ports_for_instance( [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] with excutils.save_and_reraise_exception(): [ 644.568539] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] self.force_reraise() [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise self.value [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] updated_port = self._update_port( [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] _ensure_no_port_binding_failure(port) [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] raise exception.PortBindingFailed(port_id=port['id']) [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] nova.exception.PortBindingFailed: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. [ 644.571629] env[61907]: ERROR nova.compute.manager [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] [ 644.572154] env[61907]: DEBUG nova.compute.utils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 644.572154] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.627s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.573224] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Build of instance 420a661e-33aa-43e5-b298-19a20e8ab33b was re-scheduled: Binding failed for port 384066c6-dfea-4a6e-bd95-81ae85b9e253, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 644.573671] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 644.573894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquiring lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.574040] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Acquired lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.574206] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.706890] env[61907]: INFO nova.compute.manager [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 92fa087f-a147-438b-9c30-c93baa5828dc] Took 1.03 seconds to deallocate network for instance. [ 645.107875] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.272202] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.618138] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5093f816-e5ef-44bb-84c7-384956ed893e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.627123] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e00a93-8054-41a2-95b6-8cc6f7c6385f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.663643] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf6428f-e380-4397-b559-20bfd5ed3a5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.671612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e322b2ef-70fb-4c6c-9967-37c753acbef9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.686620] env[61907]: DEBUG nova.compute.provider_tree [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.739027] env[61907]: INFO nova.scheduler.client.report [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Deleted allocations for instance 92fa087f-a147-438b-9c30-c93baa5828dc [ 645.775589] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Releasing lock "refresh_cache-420a661e-33aa-43e5-b298-19a20e8ab33b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.775897] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 645.776144] env[61907]: DEBUG nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 645.776672] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 645.809083] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.191385] env[61907]: DEBUG nova.scheduler.client.report [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 646.251716] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e51b8cf4-5c98-478f-aa97-fc11e83754e6 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "92fa087f-a147-438b-9c30-c93baa5828dc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.705s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.311783] env[61907]: DEBUG nova.network.neutron [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.695072] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.125s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.695732] env[61907]: ERROR nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] Traceback (most recent call last): [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.driver.spawn(context, instance, image_meta, [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] vm_ref = self.build_virtual_machine(instance, [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.695732] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] for vif in network_info: [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return self._sync_wrapper(fn, *args, **kwargs) [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.wait() [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self[:] = self._gt.wait() [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return self._exit_event.wait() [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] current.throw(*self._exc) [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.696051] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] result = function(*args, **kwargs) [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] return func(*args, **kwargs) [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise e [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] nwinfo = self.network_api.allocate_for_instance( [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] created_port_ids = self._update_ports_for_instance( [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] with excutils.save_and_reraise_exception(): [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] self.force_reraise() [ 646.696354] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise self.value [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] updated_port = self._update_port( [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] _ensure_no_port_binding_failure(port) [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] raise exception.PortBindingFailed(port_id=port['id']) [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] nova.exception.PortBindingFailed: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. [ 646.696704] env[61907]: ERROR nova.compute.manager [instance: a7e883a1-aab9-4372-8379-144b373062e2] [ 646.696704] env[61907]: DEBUG nova.compute.utils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.698373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.534s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.699279] env[61907]: INFO nova.compute.claims [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 646.701976] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Build of instance a7e883a1-aab9-4372-8379-144b373062e2 was re-scheduled: Binding failed for port 6a8682b1-c166-4fab-bc29-e0271db91e3a, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 646.702418] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 646.702716] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquiring lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.702858] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Acquired lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.703020] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 646.754322] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 646.814707] env[61907]: INFO nova.compute.manager [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] [instance: 420a661e-33aa-43e5-b298-19a20e8ab33b] Took 1.04 seconds to deallocate network for instance. [ 647.225054] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.232987] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.233260] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.281646] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.309452] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.814491] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Releasing lock "refresh_cache-a7e883a1-aab9-4372-8379-144b373062e2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.814755] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 647.814942] env[61907]: DEBUG nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 647.815122] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 647.831457] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.848650] env[61907]: INFO nova.scheduler.client.report [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Deleted allocations for instance 420a661e-33aa-43e5-b298-19a20e8ab33b [ 648.084020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3526131f-1a2e-4a48-99c6-b0f5b3d79507 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.091871] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cfbfb4-782b-4dc9-9ffb-44af726f6945 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.123167] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f608fa9-250f-4be1-921c-155c0d3d4fb4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.133999] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceaefa1e-9a2e-40e3-a5d5-272093d2f1b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.147377] env[61907]: DEBUG nova.compute.provider_tree [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 648.336092] env[61907]: DEBUG nova.network.neutron [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.356614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-44bba45d-9a41-47e4-af0e-165e8ba171da tempest-ImagesOneServerTestJSON-1362190004 tempest-ImagesOneServerTestJSON-1362190004-project-member] Lock "420a661e-33aa-43e5-b298-19a20e8ab33b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.793s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.655021] env[61907]: DEBUG nova.scheduler.client.report [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 648.839189] env[61907]: INFO nova.compute.manager [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] [instance: a7e883a1-aab9-4372-8379-144b373062e2] Took 1.02 seconds to deallocate network for instance. [ 648.862349] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 649.157617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 649.158185] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 649.160941] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.291s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 649.163035] env[61907]: INFO nova.compute.claims [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 649.396216] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.670093] env[61907]: DEBUG nova.compute.utils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 649.672082] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 649.672412] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 649.716251] env[61907]: DEBUG nova.policy [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b83f3e52ab544f79a2f0a6f1e2ce1de9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '70ae6743ccc04615801a4e0d0f3935a6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 649.881782] env[61907]: INFO nova.scheduler.client.report [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Deleted allocations for instance a7e883a1-aab9-4372-8379-144b373062e2 [ 650.098359] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Successfully created port: 216a83cf-1f9f-4159-bb5c-7e168dca4e85 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 650.175997] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 650.393493] env[61907]: DEBUG oslo_concurrency.lockutils [None req-69acb5a8-6067-4ae4-8ca0-53ef756216ac tempest-ServerPasswordTestJSON-479342830 tempest-ServerPasswordTestJSON-479342830-project-member] Lock "a7e883a1-aab9-4372-8379-144b373062e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.840s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.600088] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c98634-d620-443c-812b-7e709ba85a53 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.610287] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe13237-3c75-4aec-bb56-73b13937fefa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.642681] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-325876a5-7050-416f-af2c-0ee830c8d787 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.650094] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4fccb4d-2310-4dd3-bf16-09397bb57d36 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.669153] env[61907]: DEBUG nova.compute.provider_tree [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.896746] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 651.166421] env[61907]: DEBUG nova.compute.manager [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Received event network-changed-216a83cf-1f9f-4159-bb5c-7e168dca4e85 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 651.166630] env[61907]: DEBUG nova.compute.manager [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Refreshing instance network info cache due to event network-changed-216a83cf-1f9f-4159-bb5c-7e168dca4e85. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 651.166841] env[61907]: DEBUG oslo_concurrency.lockutils [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] Acquiring lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.166984] env[61907]: DEBUG oslo_concurrency.lockutils [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] Acquired lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.167162] env[61907]: DEBUG nova.network.neutron [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Refreshing network info cache for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 651.171672] env[61907]: DEBUG nova.scheduler.client.report [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 651.188526] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 651.220366] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 651.220636] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 651.220729] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 651.220955] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 651.221746] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 651.221827] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 651.222051] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 651.222238] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 651.222380] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 651.222696] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 651.222872] env[61907]: DEBUG nova.virt.hardware [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 651.223977] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05f48c2f-2ad6-4523-8c48-943e4530c48a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.234814] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d45b2bfe-d1df-4e79-99ff-2de5333570b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.351601] env[61907]: ERROR nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 651.351601] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 651.351601] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.351601] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.351601] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.351601] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.351601] env[61907]: ERROR nova.compute.manager raise self.value [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.351601] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 651.351601] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.351601] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 651.352146] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.352146] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 651.352146] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 651.352146] env[61907]: ERROR nova.compute.manager [ 651.352146] env[61907]: Traceback (most recent call last): [ 651.352146] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 651.352146] env[61907]: listener.cb(fileno) [ 651.352146] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.352146] env[61907]: result = function(*args, **kwargs) [ 651.352146] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.352146] env[61907]: return func(*args, **kwargs) [ 651.352146] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 651.352146] env[61907]: raise e [ 651.352146] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 651.352146] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 651.352146] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.352146] env[61907]: created_port_ids = self._update_ports_for_instance( [ 651.352146] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.352146] env[61907]: with excutils.save_and_reraise_exception(): [ 651.352146] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.352146] env[61907]: self.force_reraise() [ 651.352146] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.352146] env[61907]: raise self.value [ 651.352146] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.352146] env[61907]: updated_port = self._update_port( [ 651.352146] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.352146] env[61907]: _ensure_no_port_binding_failure(port) [ 651.352146] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.352146] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 651.352963] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 651.352963] env[61907]: Removing descriptor: 17 [ 651.352963] env[61907]: ERROR nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Traceback (most recent call last): [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] yield resources [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.driver.spawn(context, instance, image_meta, [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 651.352963] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] vm_ref = self.build_virtual_machine(instance, [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] for vif in network_info: [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self._sync_wrapper(fn, *args, **kwargs) [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.wait() [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self[:] = self._gt.wait() [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self._exit_event.wait() [ 651.353319] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] result = hub.switch() [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self.greenlet.switch() [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] result = function(*args, **kwargs) [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return func(*args, **kwargs) [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise e [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] nwinfo = self.network_api.allocate_for_instance( [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 651.353684] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] created_port_ids = self._update_ports_for_instance( [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] with excutils.save_and_reraise_exception(): [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.force_reraise() [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise self.value [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] updated_port = self._update_port( [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] _ensure_no_port_binding_failure(port) [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 651.354053] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise exception.PortBindingFailed(port_id=port['id']) [ 651.354481] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 651.354481] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] [ 651.354481] env[61907]: INFO nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Terminating instance [ 651.420268] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.677283] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.677816] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 651.680313] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.292s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.702119] env[61907]: DEBUG nova.network.neutron [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.856861] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquiring lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 651.890353] env[61907]: DEBUG nova.network.neutron [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.186858] env[61907]: DEBUG nova.compute.utils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 652.197963] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 652.201025] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 652.311905] env[61907]: DEBUG nova.policy [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41fc76e7ed9a46a09f96f11b28556ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b57c1194e0f14ae498d99d302da346c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 652.393780] env[61907]: DEBUG oslo_concurrency.lockutils [req-3a05edd0-f850-4aec-bd06-de7447dd57fa req-01aefd2e-1698-4efd-a3c6-5ff3460dfdf0 service nova] Releasing lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.394227] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquired lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.394407] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.574751] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72ed548-3672-4a07-8132-b3d27eeedd65 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.582119] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81bd944-b9fd-4551-b631-cf54ec99e7c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.612606] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c823429-645a-4539-8771-642bfac7e63c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.620942] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e357477-7158-4ffd-a5ce-9a41af945bf4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.635479] env[61907]: DEBUG nova.compute.provider_tree [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.697702] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 652.921263] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.980140] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Successfully created port: f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 653.135214] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.138945] env[61907]: DEBUG nova.scheduler.client.report [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 653.350164] env[61907]: DEBUG nova.compute.manager [req-9b9eb52b-aaab-4c70-910a-a228346688e2 req-ff113b01-fd7a-485c-b818-e2bd5fed21f7 service nova] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Received event network-vif-deleted-216a83cf-1f9f-4159-bb5c-7e168dca4e85 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 653.640021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Releasing lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.640021] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 653.640021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.640021] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db1347a5-8aa2-4e8f-8dab-41f0e714de11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.645253] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.965s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.646100] env[61907]: ERROR nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Traceback (most recent call last): [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.driver.spawn(context, instance, image_meta, [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] vm_ref = self.build_virtual_machine(instance, [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.646100] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] for vif in network_info: [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self._sync_wrapper(fn, *args, **kwargs) [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.wait() [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self[:] = self._gt.wait() [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self._exit_event.wait() [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] result = hub.switch() [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.646498] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return self.greenlet.switch() [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] result = function(*args, **kwargs) [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] return func(*args, **kwargs) [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise e [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] nwinfo = self.network_api.allocate_for_instance( [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] created_port_ids = self._update_ports_for_instance( [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] with excutils.save_and_reraise_exception(): [ 653.646821] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] self.force_reraise() [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise self.value [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] updated_port = self._update_port( [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] _ensure_no_port_binding_failure(port) [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] raise exception.PortBindingFailed(port_id=port['id']) [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] nova.exception.PortBindingFailed: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. [ 653.647135] env[61907]: ERROR nova.compute.manager [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] [ 653.648040] env[61907]: DEBUG nova.compute.utils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.652516] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Build of instance d95c4916-660e-4ead-ba87-2f69871d5f1e was re-scheduled: Binding failed for port 2db53abe-3598-498f-b4d7-2339c61124a6, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 653.653702] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 653.654083] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.654299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.654983] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.658851] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.105s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.660468] env[61907]: INFO nova.compute.claims [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.670557] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5010c0ef-a0ba-49a5-88ff-0c431b233bd0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.703025] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0da17f71-46ee-42f3-99f7-62f4db6e1d9d could not be found. [ 653.703025] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.703025] env[61907]: INFO nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Took 0.06 seconds to destroy the instance on the hypervisor. [ 653.703025] env[61907]: DEBUG oslo.service.loopingcall [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.703025] env[61907]: DEBUG nova.compute.manager [-] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 653.703807] env[61907]: DEBUG nova.network.neutron [-] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.707456] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.714420] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 653.720331] env[61907]: DEBUG nova.network.neutron [-] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.753753] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.753935] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.754109] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.754299] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.754440] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.754586] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.755160] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.755160] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.755160] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.755290] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.755417] env[61907]: DEBUG nova.virt.hardware [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.756290] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fab932b-3171-4bc3-ac4c-7a2d5ce8cb30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.766110] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee0b4dc-d8d4-4dc6-8988-3541aee299f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.861294] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.225506] env[61907]: DEBUG nova.network.neutron [-] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.294920] env[61907]: ERROR nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 654.294920] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.294920] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.294920] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.294920] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.294920] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.294920] env[61907]: ERROR nova.compute.manager raise self.value [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.294920] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 654.294920] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.294920] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 654.295489] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.295489] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 654.295489] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 654.295489] env[61907]: ERROR nova.compute.manager [ 654.295489] env[61907]: Traceback (most recent call last): [ 654.295489] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 654.295489] env[61907]: listener.cb(fileno) [ 654.295489] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.295489] env[61907]: result = function(*args, **kwargs) [ 654.295489] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.295489] env[61907]: return func(*args, **kwargs) [ 654.295489] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.295489] env[61907]: raise e [ 654.295489] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.295489] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 654.295489] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.295489] env[61907]: created_port_ids = self._update_ports_for_instance( [ 654.295489] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.295489] env[61907]: with excutils.save_and_reraise_exception(): [ 654.295489] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.295489] env[61907]: self.force_reraise() [ 654.295489] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.295489] env[61907]: raise self.value [ 654.295489] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.295489] env[61907]: updated_port = self._update_port( [ 654.295489] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.295489] env[61907]: _ensure_no_port_binding_failure(port) [ 654.295489] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.295489] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 654.296718] env[61907]: nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 654.296718] env[61907]: Removing descriptor: 17 [ 654.296718] env[61907]: ERROR nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Traceback (most recent call last): [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] yield resources [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.driver.spawn(context, instance, image_meta, [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self._vmops.spawn(context, instance, image_meta, injected_files, [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 654.296718] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] vm_ref = self.build_virtual_machine(instance, [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] vif_infos = vmwarevif.get_vif_info(self._session, [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] for vif in network_info: [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self._sync_wrapper(fn, *args, **kwargs) [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.wait() [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self[:] = self._gt.wait() [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self._exit_event.wait() [ 654.297073] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] result = hub.switch() [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self.greenlet.switch() [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] result = function(*args, **kwargs) [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return func(*args, **kwargs) [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise e [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] nwinfo = self.network_api.allocate_for_instance( [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 654.297647] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] created_port_ids = self._update_ports_for_instance( [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] with excutils.save_and_reraise_exception(): [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.force_reraise() [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise self.value [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] updated_port = self._update_port( [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] _ensure_no_port_binding_failure(port) [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 654.298093] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise exception.PortBindingFailed(port_id=port['id']) [ 654.298426] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 654.298426] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] [ 654.298426] env[61907]: INFO nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Terminating instance [ 654.362846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-d95c4916-660e-4ead-ba87-2f69871d5f1e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.363664] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 654.363664] env[61907]: DEBUG nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 654.363664] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.382680] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.728254] env[61907]: INFO nova.compute.manager [-] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Took 1.03 seconds to deallocate network for instance. [ 654.732828] env[61907]: DEBUG nova.compute.claims [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.733049] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.801971] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 654.802185] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.802358] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.885129] env[61907]: DEBUG nova.network.neutron [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.097272] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c51bb7-f4ea-49f9-822d-02357e61359e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.108983] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3570b0a1-c20b-4d37-9c08-4220541c296e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.142021] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2efc58e-2485-4514-bb4d-dd571c32c604 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.146827] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb640631-2b50-4378-8986-abdd0c357930 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.161448] env[61907]: DEBUG nova.compute.provider_tree [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.329809] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.390225] env[61907]: DEBUG nova.compute.manager [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Received event network-changed-f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 655.390475] env[61907]: DEBUG nova.compute.manager [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Refreshing instance network info cache due to event network-changed-f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 655.390682] env[61907]: DEBUG oslo_concurrency.lockutils [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] Acquiring lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.391485] env[61907]: INFO nova.compute.manager [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: d95c4916-660e-4ead-ba87-2f69871d5f1e] Took 1.03 seconds to deallocate network for instance. [ 655.446867] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.546434] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "a6a117b3-7cc4-43c5-8202-12514f695e48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.546434] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.668054] env[61907]: DEBUG nova.scheduler.client.report [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 655.949700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.950161] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 655.950350] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.950652] env[61907]: DEBUG oslo_concurrency.lockutils [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] Acquired lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.950821] env[61907]: DEBUG nova.network.neutron [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Refreshing network info cache for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 655.951812] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-39212c7f-cdf7-4082-9a5f-e0963cf9b304 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.964050] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2b8712-0d16-4837-9eeb-3c6b46e55ed8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.988483] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab7d4c93-fbc9-4446-ac57-f479ab408180 could not be found. [ 655.988778] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.988979] env[61907]: INFO nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.989236] env[61907]: DEBUG oslo.service.loopingcall [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.989462] env[61907]: DEBUG nova.compute.manager [-] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 655.989554] env[61907]: DEBUG nova.network.neutron [-] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.005172] env[61907]: DEBUG nova.network.neutron [-] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.178275] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.519s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.178543] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 656.181331] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.237s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.430941] env[61907]: INFO nova.scheduler.client.report [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance d95c4916-660e-4ead-ba87-2f69871d5f1e [ 656.476414] env[61907]: DEBUG nova.network.neutron [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.509022] env[61907]: DEBUG nova.network.neutron [-] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.584776] env[61907]: DEBUG nova.network.neutron [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.688314] env[61907]: DEBUG nova.compute.utils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.692801] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 656.692801] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 656.748618] env[61907]: DEBUG nova.policy [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e2baa84496ab4b99b86ca63c520e6e82', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '660238a0aa664cda9acad79d183a0c2a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 656.945195] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c4cdd83b-927f-493a-a04e-45ea3e10589e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "d95c4916-660e-4ead-ba87-2f69871d5f1e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.383s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.015012] env[61907]: INFO nova.compute.manager [-] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Took 1.02 seconds to deallocate network for instance. [ 657.019529] env[61907]: DEBUG nova.compute.claims [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.019862] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.041642] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5dac79-5f32-4b28-b3dd-057c96159d28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.052174] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17341845-6aa7-4384-8271-6b0ba5e415ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.086082] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1cf36fd-edd5-45f2-9ba9-e763d1c4a2e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.088941] env[61907]: DEBUG oslo_concurrency.lockutils [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] Releasing lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.089300] env[61907]: DEBUG nova.compute.manager [req-4ed2b4a9-3244-4101-a49e-5349379a2f42 req-bfc73817-4c12-4e63-a279-4881625eadfe service nova] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Received event network-vif-deleted-f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 657.094373] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c220a53d-9623-4597-ad28-a7535313f2b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.098949] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Successfully created port: 87d5456d-e601-4bbd-bba5-66683873c2f3 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 657.111103] env[61907]: DEBUG nova.compute.provider_tree [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 657.193853] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 657.334369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "865d386d-c5f0-439b-b50b-b613993ec9b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.334606] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.449325] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 657.616178] env[61907]: DEBUG nova.scheduler.client.report [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 657.700384] env[61907]: INFO nova.virt.block_device [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Booting with volume aa31e190-803f-464d-b8e7-504f05f564e8 at /dev/sda [ 657.756255] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e4bd873c-7d43-4b54-931e-555d9eb9a835 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.763582] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279cde61-b3dd-47e1-961e-44c45c40e20f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.790671] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f8ee633-e430-4bec-abab-61fb6b5b363f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.799752] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6469602c-1384-4d9e-8b12-fb6113f7952b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.824394] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402e9d79-35e4-4ec0-bb2b-12817c87f929 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.830950] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a92fb0c1-e0d9-4716-830f-5e62a1d39bc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.844312] env[61907]: DEBUG nova.virt.block_device [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating existing volume attachment record: 7c085b7c-1a14-4818-9143-32ad4bb759cd {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 657.977225] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.125392] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.944s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 658.126056] env[61907]: ERROR nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Traceback (most recent call last): [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.driver.spawn(context, instance, image_meta, [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] vm_ref = self.build_virtual_machine(instance, [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] vif_infos = vmwarevif.get_vif_info(self._session, [ 658.126056] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] for vif in network_info: [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self._sync_wrapper(fn, *args, **kwargs) [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.wait() [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self[:] = self._gt.wait() [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self._exit_event.wait() [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] result = hub.switch() [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 658.126398] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return self.greenlet.switch() [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] result = function(*args, **kwargs) [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] return func(*args, **kwargs) [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise e [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] nwinfo = self.network_api.allocate_for_instance( [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] created_port_ids = self._update_ports_for_instance( [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] with excutils.save_and_reraise_exception(): [ 658.126713] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] self.force_reraise() [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise self.value [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] updated_port = self._update_port( [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] _ensure_no_port_binding_failure(port) [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] raise exception.PortBindingFailed(port_id=port['id']) [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] nova.exception.PortBindingFailed: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. [ 658.127023] env[61907]: ERROR nova.compute.manager [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] [ 658.127311] env[61907]: DEBUG nova.compute.utils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 658.128588] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.903s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.130267] env[61907]: INFO nova.compute.claims [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 658.137074] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Build of instance 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe was re-scheduled: Binding failed for port d5f5e136-ba67-41a9-bce1-23cff18cb55f, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 658.137537] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 658.137762] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquiring lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.137912] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Acquired lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.138091] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.163558] env[61907]: DEBUG nova.compute.manager [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Received event network-changed-87d5456d-e601-4bbd-bba5-66683873c2f3 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 658.163755] env[61907]: DEBUG nova.compute.manager [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Refreshing instance network info cache due to event network-changed-87d5456d-e601-4bbd-bba5-66683873c2f3. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 658.163961] env[61907]: DEBUG oslo_concurrency.lockutils [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] Acquiring lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.164122] env[61907]: DEBUG oslo_concurrency.lockutils [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] Acquired lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.164282] env[61907]: DEBUG nova.network.neutron [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Refreshing network info cache for port 87d5456d-e601-4bbd-bba5-66683873c2f3 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 658.324559] env[61907]: ERROR nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 658.324559] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 658.324559] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.324559] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.324559] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.324559] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.324559] env[61907]: ERROR nova.compute.manager raise self.value [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.324559] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.324559] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.324559] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.324991] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.324991] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.324991] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 658.324991] env[61907]: ERROR nova.compute.manager [ 658.324991] env[61907]: Traceback (most recent call last): [ 658.324991] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.324991] env[61907]: listener.cb(fileno) [ 658.324991] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.324991] env[61907]: result = function(*args, **kwargs) [ 658.324991] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 658.324991] env[61907]: return func(*args, **kwargs) [ 658.324991] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 658.324991] env[61907]: raise e [ 658.324991] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 658.324991] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 658.324991] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.324991] env[61907]: created_port_ids = self._update_ports_for_instance( [ 658.324991] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.324991] env[61907]: with excutils.save_and_reraise_exception(): [ 658.324991] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.324991] env[61907]: self.force_reraise() [ 658.324991] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.324991] env[61907]: raise self.value [ 658.324991] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.324991] env[61907]: updated_port = self._update_port( [ 658.324991] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.324991] env[61907]: _ensure_no_port_binding_failure(port) [ 658.324991] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.324991] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.325785] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 658.325785] env[61907]: Removing descriptor: 17 [ 658.662186] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.685998] env[61907]: DEBUG nova.network.neutron [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.813702] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.822182] env[61907]: DEBUG nova.network.neutron [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.092038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.092256] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.317201] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Releasing lock "refresh_cache-61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.317463] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 659.317627] env[61907]: DEBUG nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 659.317798] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.324254] env[61907]: DEBUG oslo_concurrency.lockutils [req-5cab4f50-d744-4ee5-8212-1cd7c451a42a req-a643d6e6-88b7-44fa-8776-7b0f8054c9d8 service nova] Releasing lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.332591] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.482184] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b5671b-0953-47a9-9fbb-ecb69090de39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.490324] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a074c27-a0fd-4084-a14f-454317bb675b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.521119] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82d8f39-b3a3-4808-92b0-6a4b696b0ecc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.528048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc4ceb4-e832-460b-b8b4-35801eaee7c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.540776] env[61907]: DEBUG nova.compute.provider_tree [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.835218] env[61907]: DEBUG nova.network.neutron [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.939245] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 659.939848] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.940107] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.940227] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.940407] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.940548] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.940693] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.940897] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.941069] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.941797] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.941797] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.941797] env[61907]: DEBUG nova.virt.hardware [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.942618] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbbff6e-e1fc-406b-a065-400822920f20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.950918] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5066d4ae-f8c2-4754-827b-299d408d73db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.964809] env[61907]: ERROR nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Traceback (most recent call last): [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] yield resources [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.driver.spawn(context, instance, image_meta, [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] vm_ref = self.build_virtual_machine(instance, [ 659.964809] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] for vif in network_info: [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return self._sync_wrapper(fn, *args, **kwargs) [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.wait() [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self[:] = self._gt.wait() [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return self._exit_event.wait() [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.965162] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] current.throw(*self._exc) [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] result = function(*args, **kwargs) [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return func(*args, **kwargs) [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise e [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] nwinfo = self.network_api.allocate_for_instance( [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] created_port_ids = self._update_ports_for_instance( [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] with excutils.save_and_reraise_exception(): [ 659.965516] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.force_reraise() [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise self.value [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] updated_port = self._update_port( [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] _ensure_no_port_binding_failure(port) [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise exception.PortBindingFailed(port_id=port['id']) [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 659.965877] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] [ 659.965877] env[61907]: INFO nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Terminating instance [ 660.044019] env[61907]: DEBUG nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 660.190463] env[61907]: DEBUG nova.compute.manager [req-f98d60a9-9559-4f7b-8ecf-81c83d1175fd req-a112d429-e9db-4952-aa1d-6b1cab9c4f25 service nova] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Received event network-vif-deleted-87d5456d-e601-4bbd-bba5-66683873c2f3 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 660.337874] env[61907]: INFO nova.compute.manager [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] [instance: 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe] Took 1.02 seconds to deallocate network for instance. [ 660.469048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquiring lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.469263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquired lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.469442] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 660.548994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.549545] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 660.551944] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 17.982s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.987103] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.054717] env[61907]: DEBUG nova.compute.utils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 661.058832] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 661.058998] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 661.070553] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.244194] env[61907]: DEBUG nova.policy [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9115523810b84ab3a4e0f822709a8eaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc8646a92534c53ba7b3335eb1ab9e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.373244] env[61907]: INFO nova.scheduler.client.report [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Deleted allocations for instance 61e030b3-7f1c-40a3-99d4-e5e9cf41cafe [ 661.518597] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Successfully created port: 0f90d605-9722-4138-9823-c17442559052 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.560162] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 661.572302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Releasing lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.572865] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 661.573302] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7025e4db-76b6-4de6-938f-971ff85558d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.582350] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f87d14-fb76-4748-baa5-3c1d58cc7a8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.594764] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 0da17f71-46ee-42f3-99f7-62f4db6e1d9d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.594988] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ab7d4c93-fbc9-4446-ac57-f479ab408180 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.595198] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 78f3d457-d76b-4a13-9f12-264e53b768b4 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.595326] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 90a4f8bb-59bd-499f-b2fc-ebce55c5f803 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 661.608318] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 78f3d457-d76b-4a13-9f12-264e53b768b4 could not be found. [ 661.608542] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 661.608875] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0b623022-b7fc-4159-8daa-8258f214c21a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.616563] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bdec2b8-2ec4-4ee0-b4d0-9a36fa5af28c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.637695] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 78f3d457-d76b-4a13-9f12-264e53b768b4 could not be found. [ 661.637913] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 661.638107] env[61907]: INFO nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Took 0.07 seconds to destroy the instance on the hypervisor. [ 661.638347] env[61907]: DEBUG oslo.service.loopingcall [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 661.639190] env[61907]: DEBUG nova.compute.manager [-] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 661.643179] env[61907]: DEBUG nova.network.neutron [-] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 661.661551] env[61907]: DEBUG nova.network.neutron [-] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 661.884798] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e468e642-87ab-4115-b882-bd4bb09f29e1 tempest-ServersAdminTestJSON-1624259992 tempest-ServersAdminTestJSON-1624259992-project-member] Lock "61e030b3-7f1c-40a3-99d4-e5e9cf41cafe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.974s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.100094] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 6ac3cdf2-eb93-4403-a514-81d56baf544e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.164882] env[61907]: DEBUG nova.network.neutron [-] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.263469] env[61907]: DEBUG nova.compute.manager [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Received event network-changed-0f90d605-9722-4138-9823-c17442559052 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 662.263622] env[61907]: DEBUG nova.compute.manager [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Refreshing instance network info cache due to event network-changed-0f90d605-9722-4138-9823-c17442559052. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 662.263833] env[61907]: DEBUG oslo_concurrency.lockutils [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] Acquiring lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.263975] env[61907]: DEBUG oslo_concurrency.lockutils [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] Acquired lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.264157] env[61907]: DEBUG nova.network.neutron [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Refreshing network info cache for port 0f90d605-9722-4138-9823-c17442559052 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 662.385374] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 662.471025] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 662.471025] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.471025] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.471025] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.471025] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.471025] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.471025] env[61907]: ERROR nova.compute.manager raise self.value [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.471025] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.471025] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.471025] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.471500] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.471500] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.471500] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 662.471500] env[61907]: ERROR nova.compute.manager [ 662.471500] env[61907]: Traceback (most recent call last): [ 662.471500] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.471500] env[61907]: listener.cb(fileno) [ 662.471500] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.471500] env[61907]: result = function(*args, **kwargs) [ 662.471500] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.471500] env[61907]: return func(*args, **kwargs) [ 662.471500] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.471500] env[61907]: raise e [ 662.471500] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.471500] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 662.471500] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.471500] env[61907]: created_port_ids = self._update_ports_for_instance( [ 662.471500] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.471500] env[61907]: with excutils.save_and_reraise_exception(): [ 662.471500] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.471500] env[61907]: self.force_reraise() [ 662.471500] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.471500] env[61907]: raise self.value [ 662.471500] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.471500] env[61907]: updated_port = self._update_port( [ 662.471500] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.471500] env[61907]: _ensure_no_port_binding_failure(port) [ 662.471500] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.471500] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.472343] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 662.472343] env[61907]: Removing descriptor: 17 [ 662.572393] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 662.597579] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.597827] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.597988] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.598198] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.598347] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.598491] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.598719] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.598898] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.599086] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.599251] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.599430] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.600554] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ab7e3c4-15dd-4f48-bdd7-8784714acdac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.607731] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 8456ccc3-aa32-4388-bf9e-2608c783405f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.609694] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333fd4ce-5aa8-4599-be6b-b92840e536ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.625308] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Traceback (most recent call last): [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] yield resources [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.driver.spawn(context, instance, image_meta, [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] vm_ref = self.build_virtual_machine(instance, [ 662.625308] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] for vif in network_info: [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return self._sync_wrapper(fn, *args, **kwargs) [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.wait() [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self[:] = self._gt.wait() [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return self._exit_event.wait() [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.625733] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] current.throw(*self._exc) [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] result = function(*args, **kwargs) [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return func(*args, **kwargs) [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise e [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] nwinfo = self.network_api.allocate_for_instance( [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] created_port_ids = self._update_ports_for_instance( [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] with excutils.save_and_reraise_exception(): [ 662.626116] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.force_reraise() [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise self.value [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] updated_port = self._update_port( [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] _ensure_no_port_binding_failure(port) [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise exception.PortBindingFailed(port_id=port['id']) [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 662.626530] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] [ 662.626530] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Terminating instance [ 662.666983] env[61907]: INFO nova.compute.manager [-] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Took 1.02 seconds to deallocate network for instance. [ 662.782363] env[61907]: DEBUG nova.network.neutron [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.880683] env[61907]: DEBUG nova.network.neutron [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.906121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.115668] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 520e2134-3553-44c2-98f5-ecd5be642d86 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.129829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.222054] env[61907]: INFO nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Took 0.55 seconds to detach 1 volumes for instance. [ 663.224890] env[61907]: DEBUG nova.compute.claims [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 663.225729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.383938] env[61907]: DEBUG oslo_concurrency.lockutils [req-256fa42a-9d87-4e91-b300-3a38dbe9e75e req-f91b70a8-6e8d-4bc3-831b-60d79ed2bb89 service nova] Releasing lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.384118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.384292] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.620960] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a1189e41-5922-4623-bb03-394d98159193 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.920449] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.024929] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.123502] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.289272] env[61907]: DEBUG nova.compute.manager [req-554ca040-9aef-4fa1-9dd3-7d79b01a79c4 req-823a9f48-e14b-4551-9cfa-85a12410fd3c service nova] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Received event network-vif-deleted-0f90d605-9722-4138-9823-c17442559052 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 664.527703] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.528229] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 664.528467] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.528780] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-065cdc55-ef76-4f84-87f7-778993394e08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.537701] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c7e246a-00ea-455c-b5fb-9da0812fdc26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.559242] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 90a4f8bb-59bd-499f-b2fc-ebce55c5f803 could not be found. [ 664.559428] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.559610] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Took 0.03 seconds to destroy the instance on the hypervisor. [ 664.559953] env[61907]: DEBUG oslo.service.loopingcall [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.560197] env[61907]: DEBUG nova.compute.manager [-] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 664.560296] env[61907]: DEBUG nova.network.neutron [-] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.575446] env[61907]: DEBUG nova.network.neutron [-] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.627091] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.077718] env[61907]: DEBUG nova.network.neutron [-] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.130042] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.580755] env[61907]: INFO nova.compute.manager [-] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Took 1.02 seconds to deallocate network for instance. [ 665.583186] env[61907]: DEBUG nova.compute.claims [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.583362] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.633618] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ee695d5f-1f9d-49b3-b5fa-180ea953e5ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.137456] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 09d97063-d94f-439a-a811-b3039de1f4d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.639390] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance af575382-25f6-475a-8803-48714fa0a37c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.142019] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.644872] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.148484] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.651579] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ea6d2b19-be25-4da3-b8af-11c3dca04df8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.154808] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance da7ddc92-f3e3-4b4d-a14b-13fd25a72314 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.660176] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f517b2e4-1a83-4519-8a5b-65d489d16b6d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.163569] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.667563] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 215d0f35-5d19-4db4-996d-571e2970e19c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.171506] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 45c58b31-d0cc-4b36-b744-448a38e11266 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.674869] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.177838] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 9209400d-d5b8-4b96-9022-f8a4c7a30471 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.681221] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a6a117b3-7cc4-43c5-8202-12514f695e48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.184279] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 865d386d-c5f0-439b-b50b-b613993ec9b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.687812] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 3c9ee76a-464d-437e-9b73-862f1da24f7e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.688173] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 673.688339] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 674.045217] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11c1cf4-4685-4ae2-9780-a716ed43b318 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.053319] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f74149d-1f9c-4a34-bfba-5142e398423a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.083675] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40403d6d-9046-4eae-b13b-0309e9cc313b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.090748] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bbfed4-1216-4cdd-97d6-ea4dded27412 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.103494] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.607032] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 675.112105] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 675.112380] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.560s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.112666] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.870s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.114187] env[61907]: INFO nova.compute.claims [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.419351] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ed76f5-9c4f-488d-ab78-a4bfb60c9ca5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.426978] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7507bae-a94b-425e-b9b7-1e6548bfbda6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.456103] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a686790-168a-4ae3-b281-30164ec526f3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.462615] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89579e07-77bd-4d4d-a28c-7047f4c50571 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.474945] env[61907]: DEBUG nova.compute.provider_tree [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.979087] env[61907]: DEBUG nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 677.484180] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.484743] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 677.488020] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.206s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.488822] env[61907]: INFO nova.compute.claims [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.993123] env[61907]: DEBUG nova.compute.utils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 677.996143] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 677.997318] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 678.045681] env[61907]: DEBUG nova.policy [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9115523810b84ab3a4e0f822709a8eaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc8646a92534c53ba7b3335eb1ab9e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 678.317352] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Successfully created port: 0164397e-bb5a-4c6e-8730-ec31fbda35ba {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 678.500359] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 678.819788] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8418f9f-3208-4e01-9671-599f19f89cbc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.827724] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a45f19-de81-4b92-bca7-c07ebec3f65e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.863702] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8645b040-4be2-456c-9c20-58e8adf3a55d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.871519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d9e6be-6231-4e69-bf77-0a388cc885b7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.885075] env[61907]: DEBUG nova.compute.provider_tree [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 679.145513] env[61907]: DEBUG nova.compute.manager [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Received event network-changed-0164397e-bb5a-4c6e-8730-ec31fbda35ba {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 679.145732] env[61907]: DEBUG nova.compute.manager [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Refreshing instance network info cache due to event network-changed-0164397e-bb5a-4c6e-8730-ec31fbda35ba. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 679.145946] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] Acquiring lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.146107] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] Acquired lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.146310] env[61907]: DEBUG nova.network.neutron [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Refreshing network info cache for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 679.301914] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 679.301914] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.301914] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.301914] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.301914] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.301914] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.301914] env[61907]: ERROR nova.compute.manager raise self.value [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.301914] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 679.301914] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.301914] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 679.302417] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.302417] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 679.302417] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 679.302417] env[61907]: ERROR nova.compute.manager [ 679.302417] env[61907]: Traceback (most recent call last): [ 679.302417] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 679.302417] env[61907]: listener.cb(fileno) [ 679.302417] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.302417] env[61907]: result = function(*args, **kwargs) [ 679.302417] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.302417] env[61907]: return func(*args, **kwargs) [ 679.302417] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 679.302417] env[61907]: raise e [ 679.302417] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.302417] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 679.302417] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.302417] env[61907]: created_port_ids = self._update_ports_for_instance( [ 679.302417] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.302417] env[61907]: with excutils.save_and_reraise_exception(): [ 679.302417] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.302417] env[61907]: self.force_reraise() [ 679.302417] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.302417] env[61907]: raise self.value [ 679.302417] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.302417] env[61907]: updated_port = self._update_port( [ 679.302417] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.302417] env[61907]: _ensure_no_port_binding_failure(port) [ 679.302417] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.302417] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 679.303262] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 679.303262] env[61907]: Removing descriptor: 17 [ 679.390608] env[61907]: DEBUG nova.scheduler.client.report [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 679.512108] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 679.537574] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 679.537838] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 679.538134] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 679.538179] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 679.538327] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 679.538473] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 679.538682] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 679.538842] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 679.539049] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 679.539237] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 679.539413] env[61907]: DEBUG nova.virt.hardware [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 679.540328] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01980562-9ca7-41b5-a40a-9d4b51d0f007 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.550038] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea559bc1-83b4-436f-bb5d-e05d0c9b4cb7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.562182] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Traceback (most recent call last): [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] yield resources [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.driver.spawn(context, instance, image_meta, [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] vm_ref = self.build_virtual_machine(instance, [ 679.562182] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] for vif in network_info: [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return self._sync_wrapper(fn, *args, **kwargs) [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.wait() [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self[:] = self._gt.wait() [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return self._exit_event.wait() [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.562501] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] current.throw(*self._exc) [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] result = function(*args, **kwargs) [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return func(*args, **kwargs) [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise e [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] nwinfo = self.network_api.allocate_for_instance( [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] created_port_ids = self._update_ports_for_instance( [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] with excutils.save_and_reraise_exception(): [ 679.562867] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.force_reraise() [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise self.value [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] updated_port = self._update_port( [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] _ensure_no_port_binding_failure(port) [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise exception.PortBindingFailed(port_id=port['id']) [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 679.563198] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] [ 679.563198] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Terminating instance [ 679.663233] env[61907]: DEBUG nova.network.neutron [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.744528] env[61907]: DEBUG nova.network.neutron [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.895728] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.408s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.896326] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 679.900671] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.503s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.903012] env[61907]: INFO nova.compute.claims [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.067612] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.246966] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ba36180-fde0-41ff-9830-70e9ca8b020e req-477b9f11-28c1-4f04-a196-c85578b498db service nova] Releasing lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.247418] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.247638] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 680.409729] env[61907]: DEBUG nova.compute.utils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.411099] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 680.411267] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.449828] env[61907]: DEBUG nova.policy [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8682003dee14f25ba212739afd3d335', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6a88771a8f84196ae271dbd57590c15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.739119] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Successfully created port: c1ba743f-2581-4733-9fb3-5f239b4b4609 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.791059] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.918583] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 680.935680] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.195821] env[61907]: DEBUG nova.compute.manager [req-60eea46a-be3b-41fd-80c6-c32c18e36604 req-66c4b3ca-0c6d-49a1-9ba0-c7e5c89ac4e5 service nova] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Received event network-vif-deleted-0164397e-bb5a-4c6e-8730-ec31fbda35ba {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 681.314187] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb2a8f0e-e58a-4dd2-b93d-fe9b3cfc1e8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.320632] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f991bca7-20be-47ce-b465-f21ce65f4450 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.351980] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612013fd-5a37-4446-bf92-1cd4fab35894 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.359362] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de48cde6-3099-4fd1-90fc-1d3e258f50d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.374227] env[61907]: DEBUG nova.compute.provider_tree [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.437709] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.437857] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 681.438932] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 681.438932] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b6379a4-b0ec-4d57-a3c0-74cf1bb23eba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.450316] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eed8b19b-f684-4bfe-87ab-fc380dd72908 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.474629] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ac3cdf2-eb93-4403-a514-81d56baf544e could not be found. [ 681.474956] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 681.475054] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 681.475313] env[61907]: DEBUG oslo.service.loopingcall [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 681.479028] env[61907]: DEBUG nova.compute.manager [-] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 681.479028] env[61907]: DEBUG nova.network.neutron [-] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 681.492456] env[61907]: DEBUG nova.network.neutron [-] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 681.842419] env[61907]: ERROR nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 681.842419] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 681.842419] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.842419] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.842419] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.842419] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.842419] env[61907]: ERROR nova.compute.manager raise self.value [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.842419] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 681.842419] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.842419] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 681.843105] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.843105] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 681.843105] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 681.843105] env[61907]: ERROR nova.compute.manager [ 681.843105] env[61907]: Traceback (most recent call last): [ 681.843105] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 681.843105] env[61907]: listener.cb(fileno) [ 681.843105] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.843105] env[61907]: result = function(*args, **kwargs) [ 681.843105] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.843105] env[61907]: return func(*args, **kwargs) [ 681.843105] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 681.843105] env[61907]: raise e [ 681.843105] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 681.843105] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 681.843105] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.843105] env[61907]: created_port_ids = self._update_ports_for_instance( [ 681.843105] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.843105] env[61907]: with excutils.save_and_reraise_exception(): [ 681.843105] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.843105] env[61907]: self.force_reraise() [ 681.843105] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.843105] env[61907]: raise self.value [ 681.843105] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.843105] env[61907]: updated_port = self._update_port( [ 681.843105] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.843105] env[61907]: _ensure_no_port_binding_failure(port) [ 681.843105] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.843105] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 681.843825] env[61907]: nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 681.843825] env[61907]: Removing descriptor: 17 [ 681.877843] env[61907]: DEBUG nova.scheduler.client.report [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 681.927348] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 681.953369] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.953738] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.953900] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.954098] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.954254] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.954401] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.954606] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.954764] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.954926] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.955099] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.955274] env[61907]: DEBUG nova.virt.hardware [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.956213] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcad68c-78a1-4d2f-a976-f44a4ad8688a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.964797] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0a1d8e3-e743-4518-9192-7b164964747b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.978403] env[61907]: ERROR nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Traceback (most recent call last): [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] yield resources [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.driver.spawn(context, instance, image_meta, [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] vm_ref = self.build_virtual_machine(instance, [ 681.978403] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] vif_infos = vmwarevif.get_vif_info(self._session, [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] for vif in network_info: [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return self._sync_wrapper(fn, *args, **kwargs) [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.wait() [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self[:] = self._gt.wait() [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return self._exit_event.wait() [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 681.978752] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] current.throw(*self._exc) [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] result = function(*args, **kwargs) [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return func(*args, **kwargs) [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise e [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] nwinfo = self.network_api.allocate_for_instance( [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] created_port_ids = self._update_ports_for_instance( [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] with excutils.save_and_reraise_exception(): [ 681.979221] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.force_reraise() [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise self.value [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] updated_port = self._update_port( [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] _ensure_no_port_binding_failure(port) [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise exception.PortBindingFailed(port_id=port['id']) [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 681.979603] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] [ 681.979603] env[61907]: INFO nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Terminating instance [ 681.994629] env[61907]: DEBUG nova.network.neutron [-] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.382994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.383587] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 682.386371] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.966s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.387885] env[61907]: INFO nova.compute.claims [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.482681] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.482881] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.483102] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 682.496579] env[61907]: INFO nova.compute.manager [-] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Took 1.02 seconds to deallocate network for instance. [ 682.498648] env[61907]: DEBUG nova.compute.claims [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 682.498821] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.892469] env[61907]: DEBUG nova.compute.utils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.897879] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 682.898068] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.948122] env[61907]: DEBUG nova.policy [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 683.012950] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.213280] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.251948] env[61907]: DEBUG nova.compute.manager [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Received event network-changed-c1ba743f-2581-4733-9fb3-5f239b4b4609 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 683.251948] env[61907]: DEBUG nova.compute.manager [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Refreshing instance network info cache due to event network-changed-c1ba743f-2581-4733-9fb3-5f239b4b4609. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 683.251948] env[61907]: DEBUG oslo_concurrency.lockutils [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] Acquiring lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.390619] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Successfully created port: c87a7e7d-7c2f-41c5-bcf5-e0951999a913 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.398235] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 683.663355] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.663760] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 683.663952] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 683.664477] env[61907]: DEBUG oslo_concurrency.lockutils [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] Acquired lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.664653] env[61907]: DEBUG nova.network.neutron [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Refreshing network info cache for port c1ba743f-2581-4733-9fb3-5f239b4b4609 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 683.665662] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba78ea68-611b-4891-a0e3-ae3aded161ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.676149] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0645c4-37dd-419b-b436-fcf4b9688e90 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.700381] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8456ccc3-aa32-4388-bf9e-2608c783405f could not be found. [ 683.700571] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 683.700727] env[61907]: INFO nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 683.700958] env[61907]: DEBUG oslo.service.loopingcall [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.703327] env[61907]: DEBUG nova.compute.manager [-] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 683.703430] env[61907]: DEBUG nova.network.neutron [-] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 683.721824] env[61907]: DEBUG nova.network.neutron [-] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 683.795151] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca28217-2022-4f5c-9778-1e3c49e007aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.805039] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ea65770-9335-4228-bfa4-aee0ccea9c98 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.838510] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d192439-7d59-4a1f-a944-33c404dc0000 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.845884] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92feec88-b441-4b9f-85a1-95c5372fcd0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.858776] env[61907]: DEBUG nova.compute.provider_tree [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.190959] env[61907]: DEBUG nova.network.neutron [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.225575] env[61907]: DEBUG nova.network.neutron [-] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.353765] env[61907]: DEBUG nova.network.neutron [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.364922] env[61907]: DEBUG nova.scheduler.client.report [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 684.407377] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 684.441838] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.442110] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.442287] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.442473] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.442620] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.442765] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.442965] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.443207] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.443463] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.443677] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.443880] env[61907]: DEBUG nova.virt.hardware [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.445113] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448bbe18-1d45-4b16-ab60-7eb9616bf9be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.448590] env[61907]: ERROR nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 684.448590] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.448590] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.448590] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.448590] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.448590] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.448590] env[61907]: ERROR nova.compute.manager raise self.value [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.448590] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.448590] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.448590] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.449119] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.449119] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.449119] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 684.449119] env[61907]: ERROR nova.compute.manager [ 684.449119] env[61907]: Traceback (most recent call last): [ 684.449119] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.449119] env[61907]: listener.cb(fileno) [ 684.449119] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.449119] env[61907]: result = function(*args, **kwargs) [ 684.449119] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.449119] env[61907]: return func(*args, **kwargs) [ 684.449119] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.449119] env[61907]: raise e [ 684.449119] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.449119] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 684.449119] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.449119] env[61907]: created_port_ids = self._update_ports_for_instance( [ 684.449119] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.449119] env[61907]: with excutils.save_and_reraise_exception(): [ 684.449119] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.449119] env[61907]: self.force_reraise() [ 684.449119] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.449119] env[61907]: raise self.value [ 684.449119] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.449119] env[61907]: updated_port = self._update_port( [ 684.449119] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.449119] env[61907]: _ensure_no_port_binding_failure(port) [ 684.449119] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.449119] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.449966] env[61907]: nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 684.449966] env[61907]: Removing descriptor: 17 [ 684.452755] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b65c0f-1366-44e7-91a8-1405f5cd7b73 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.466194] env[61907]: ERROR nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Traceback (most recent call last): [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] yield resources [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.driver.spawn(context, instance, image_meta, [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] vm_ref = self.build_virtual_machine(instance, [ 684.466194] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] for vif in network_info: [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return self._sync_wrapper(fn, *args, **kwargs) [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.wait() [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self[:] = self._gt.wait() [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return self._exit_event.wait() [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 684.466591] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] current.throw(*self._exc) [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] result = function(*args, **kwargs) [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return func(*args, **kwargs) [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise e [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] nwinfo = self.network_api.allocate_for_instance( [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] created_port_ids = self._update_ports_for_instance( [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] with excutils.save_and_reraise_exception(): [ 684.467017] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.force_reraise() [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise self.value [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] updated_port = self._update_port( [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] _ensure_no_port_binding_failure(port) [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise exception.PortBindingFailed(port_id=port['id']) [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 684.467410] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] [ 684.467410] env[61907]: INFO nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Terminating instance [ 684.727849] env[61907]: INFO nova.compute.manager [-] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Took 1.02 seconds to deallocate network for instance. [ 684.731420] env[61907]: DEBUG nova.compute.claims [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 684.731420] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.857144] env[61907]: DEBUG oslo_concurrency.lockutils [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] Releasing lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.857576] env[61907]: DEBUG nova.compute.manager [req-af97fbc3-2da6-496d-9320-1c336a8f3bb4 req-72a664c4-27fd-4e95-af88-04d38b82e7e2 service nova] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Received event network-vif-deleted-c1ba743f-2581-4733-9fb3-5f239b4b4609 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 684.869819] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.870322] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 684.873036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.140s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.970790] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.970979] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.971174] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.303664] env[61907]: DEBUG nova.compute.manager [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Received event network-changed-c87a7e7d-7c2f-41c5-bcf5-e0951999a913 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 685.303664] env[61907]: DEBUG nova.compute.manager [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Refreshing instance network info cache due to event network-changed-c87a7e7d-7c2f-41c5-bcf5-e0951999a913. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 685.303664] env[61907]: DEBUG oslo_concurrency.lockutils [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] Acquiring lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.377811] env[61907]: DEBUG nova.compute.utils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.383405] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 685.383405] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 685.450966] env[61907]: DEBUG nova.policy [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '05978afa12534e0dbbfce7d9e8741f4d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b07213d2549f4066a055c7048950e0b6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.504170] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.634015] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.711308] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97aeca31-9ac5-425d-ab93-32ff686968e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.719549] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83d0dae7-84be-4cdd-9cb0-88cf82200b9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.749109] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a659bdc-5e26-4e53-80e1-2621e0c0fe42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.756776] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0498323-7f92-4dcf-9782-68fc3c0695f3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.769890] env[61907]: DEBUG nova.compute.provider_tree [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.885603] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 686.014622] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Successfully created port: cba547b3-4808-4176-bf9e-a581c2c6c6c0 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.136305] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.136768] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 686.136972] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.137355] env[61907]: DEBUG oslo_concurrency.lockutils [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] Acquired lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.137530] env[61907]: DEBUG nova.network.neutron [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Refreshing network info cache for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 686.139625] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6294f9f-b7a1-4fc5-910b-b6cc501bd41d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.151018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68b6a126-ef73-43a5-9414-68044fdb387a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.171812] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 520e2134-3553-44c2-98f5-ecd5be642d86 could not be found. [ 686.172024] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.172210] env[61907]: INFO nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Took 0.04 seconds to destroy the instance on the hypervisor. [ 686.172440] env[61907]: DEBUG oslo.service.loopingcall [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.172642] env[61907]: DEBUG nova.compute.manager [-] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 686.172732] env[61907]: DEBUG nova.network.neutron [-] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.192390] env[61907]: DEBUG nova.network.neutron [-] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.277195] env[61907]: DEBUG nova.scheduler.client.report [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 686.658168] env[61907]: DEBUG nova.network.neutron [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.694855] env[61907]: DEBUG nova.network.neutron [-] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.770501] env[61907]: DEBUG nova.network.neutron [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.784022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.784668] env[61907]: ERROR nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Traceback (most recent call last): [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.driver.spawn(context, instance, image_meta, [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] vm_ref = self.build_virtual_machine(instance, [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.784668] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] for vif in network_info: [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self._sync_wrapper(fn, *args, **kwargs) [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.wait() [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self[:] = self._gt.wait() [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self._exit_event.wait() [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] result = hub.switch() [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.785068] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return self.greenlet.switch() [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] result = function(*args, **kwargs) [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] return func(*args, **kwargs) [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise e [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] nwinfo = self.network_api.allocate_for_instance( [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] created_port_ids = self._update_ports_for_instance( [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] with excutils.save_and_reraise_exception(): [ 686.785503] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] self.force_reraise() [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise self.value [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] updated_port = self._update_port( [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] _ensure_no_port_binding_failure(port) [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] raise exception.PortBindingFailed(port_id=port['id']) [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] nova.exception.PortBindingFailed: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. [ 686.785936] env[61907]: ERROR nova.compute.manager [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] [ 686.786413] env[61907]: DEBUG nova.compute.utils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.786758] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Build of instance 0da17f71-46ee-42f3-99f7-62f4db6e1d9d was re-scheduled: Binding failed for port 216a83cf-1f9f-4159-bb5c-7e168dca4e85, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 686.787222] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 686.787455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquiring lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.787601] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Acquired lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.787757] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.789656] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.769s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.893174] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 686.919079] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 686.919079] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 686.919079] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 686.919297] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 686.919297] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 686.919297] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 686.919297] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 686.919297] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 686.919468] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 686.919468] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 686.919468] env[61907]: DEBUG nova.virt.hardware [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.920146] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809680b6-57db-4128-b4d9-5fe9faf294ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.928597] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2ecb3d-cb36-4cff-9661-2aa25b98082e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.039185] env[61907]: ERROR nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 687.039185] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.039185] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.039185] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.039185] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.039185] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.039185] env[61907]: ERROR nova.compute.manager raise self.value [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.039185] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 687.039185] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.039185] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 687.039620] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.039620] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 687.039620] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 687.039620] env[61907]: ERROR nova.compute.manager [ 687.039620] env[61907]: Traceback (most recent call last): [ 687.039620] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 687.039620] env[61907]: listener.cb(fileno) [ 687.039620] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.039620] env[61907]: result = function(*args, **kwargs) [ 687.039620] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.039620] env[61907]: return func(*args, **kwargs) [ 687.039620] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.039620] env[61907]: raise e [ 687.039620] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.039620] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 687.039620] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.039620] env[61907]: created_port_ids = self._update_ports_for_instance( [ 687.039620] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.039620] env[61907]: with excutils.save_and_reraise_exception(): [ 687.039620] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.039620] env[61907]: self.force_reraise() [ 687.039620] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.039620] env[61907]: raise self.value [ 687.039620] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.039620] env[61907]: updated_port = self._update_port( [ 687.039620] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.039620] env[61907]: _ensure_no_port_binding_failure(port) [ 687.039620] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.039620] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 687.040329] env[61907]: nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 687.040329] env[61907]: Removing descriptor: 17 [ 687.040329] env[61907]: ERROR nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] Traceback (most recent call last): [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] yield resources [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.driver.spawn(context, instance, image_meta, [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.040329] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] vm_ref = self.build_virtual_machine(instance, [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] for vif in network_info: [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self._sync_wrapper(fn, *args, **kwargs) [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.wait() [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self[:] = self._gt.wait() [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self._exit_event.wait() [ 687.040631] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] result = hub.switch() [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self.greenlet.switch() [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] result = function(*args, **kwargs) [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return func(*args, **kwargs) [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise e [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] nwinfo = self.network_api.allocate_for_instance( [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 687.040955] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] created_port_ids = self._update_ports_for_instance( [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] with excutils.save_and_reraise_exception(): [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.force_reraise() [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise self.value [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] updated_port = self._update_port( [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] _ensure_no_port_binding_failure(port) [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.041303] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise exception.PortBindingFailed(port_id=port['id']) [ 687.041627] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 687.041627] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] [ 687.041627] env[61907]: INFO nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Terminating instance [ 687.197374] env[61907]: INFO nova.compute.manager [-] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Took 1.02 seconds to deallocate network for instance. [ 687.200155] env[61907]: DEBUG nova.compute.claims [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.200360] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.273281] env[61907]: DEBUG oslo_concurrency.lockutils [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] Releasing lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.273490] env[61907]: DEBUG nova.compute.manager [req-ada0fa50-a25e-4115-aff5-35c2c228ca60 req-600e0875-c3ce-4a7c-9643-9c2e42d899d1 service nova] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Received event network-vif-deleted-c87a7e7d-7c2f-41c5-bcf5-e0951999a913 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 687.313906] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.340842] env[61907]: DEBUG nova.compute.manager [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Received event network-changed-cba547b3-4808-4176-bf9e-a581c2c6c6c0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 687.341228] env[61907]: DEBUG nova.compute.manager [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Refreshing instance network info cache due to event network-changed-cba547b3-4808-4176-bf9e-a581c2c6c6c0. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 687.341458] env[61907]: DEBUG oslo_concurrency.lockutils [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] Acquiring lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.341598] env[61907]: DEBUG oslo_concurrency.lockutils [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] Acquired lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.341756] env[61907]: DEBUG nova.network.neutron [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Refreshing network info cache for port cba547b3-4808-4176-bf9e-a581c2c6c6c0 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 687.417347] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.545720] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquiring lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.623260] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eba4622-f38f-4e1d-b242-1dcaf552209a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.631373] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd2f887c-7393-4d7a-97ce-e164c306d377 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.661546] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4418181a-aa31-4c49-b649-4538a13a93fc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.668549] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15077ca5-383f-4cea-a39f-c085a444099b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.681606] env[61907]: DEBUG nova.compute.provider_tree [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.861515] env[61907]: DEBUG nova.network.neutron [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.921811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Releasing lock "refresh_cache-0da17f71-46ee-42f3-99f7-62f4db6e1d9d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.921811] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 687.921811] env[61907]: DEBUG nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 687.921811] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.943903] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.951611] env[61907]: DEBUG nova.network.neutron [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.184533] env[61907]: DEBUG nova.scheduler.client.report [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 688.446824] env[61907]: DEBUG nova.network.neutron [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.454918] env[61907]: DEBUG oslo_concurrency.lockutils [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] Releasing lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.454918] env[61907]: DEBUG nova.compute.manager [req-c5285d33-a31a-4207-986a-dce6dde9bbb5 req-ddfcd023-2702-41ee-a645-6eb8ab1cbee7 service nova] [instance: a1189e41-5922-4623-bb03-394d98159193] Received event network-vif-deleted-cba547b3-4808-4176-bf9e-a581c2c6c6c0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 688.454918] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquired lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.454918] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.689228] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.689877] env[61907]: ERROR nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Traceback (most recent call last): [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.driver.spawn(context, instance, image_meta, [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] vm_ref = self.build_virtual_machine(instance, [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.689877] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] for vif in network_info: [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self._sync_wrapper(fn, *args, **kwargs) [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.wait() [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self[:] = self._gt.wait() [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self._exit_event.wait() [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] result = hub.switch() [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 688.690255] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return self.greenlet.switch() [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] result = function(*args, **kwargs) [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] return func(*args, **kwargs) [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise e [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] nwinfo = self.network_api.allocate_for_instance( [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] created_port_ids = self._update_ports_for_instance( [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] with excutils.save_and_reraise_exception(): [ 688.690773] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] self.force_reraise() [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise self.value [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] updated_port = self._update_port( [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] _ensure_no_port_binding_failure(port) [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] raise exception.PortBindingFailed(port_id=port['id']) [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] nova.exception.PortBindingFailed: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. [ 688.691169] env[61907]: ERROR nova.compute.manager [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] [ 688.691487] env[61907]: DEBUG nova.compute.utils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.692266] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Build of instance ab7d4c93-fbc9-4446-ac57-f479ab408180 was re-scheduled: Binding failed for port f7b735f4-ce6c-4c1e-ae47-a106baa7cfb7, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 688.692689] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 688.692911] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.693069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.693229] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.694235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.717s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.696122] env[61907]: INFO nova.compute.claims [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.949666] env[61907]: INFO nova.compute.manager [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] [instance: 0da17f71-46ee-42f3-99f7-62f4db6e1d9d] Took 1.03 seconds to deallocate network for instance. [ 688.976931] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.056075] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.217612] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.309686] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.558982] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Releasing lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.559591] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 689.559821] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 689.560170] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5e661ca2-4b3b-43ad-9733-7f7284c1c592 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.568917] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31819e2f-98c8-4961-bc5b-6e7399bfb7b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.589928] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1189e41-5922-4623-bb03-394d98159193 could not be found. [ 689.590196] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 689.590396] env[61907]: INFO nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Took 0.03 seconds to destroy the instance on the hypervisor. [ 689.590639] env[61907]: DEBUG oslo.service.loopingcall [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 689.590853] env[61907]: DEBUG nova.compute.manager [-] [instance: a1189e41-5922-4623-bb03-394d98159193] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 689.590948] env[61907]: DEBUG nova.network.neutron [-] [instance: a1189e41-5922-4623-bb03-394d98159193] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.605063] env[61907]: DEBUG nova.network.neutron [-] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.812485] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-ab7d4c93-fbc9-4446-ac57-f479ab408180" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.812774] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 689.812897] env[61907]: DEBUG nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 689.813074] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.827174] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.975564] env[61907]: INFO nova.scheduler.client.report [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Deleted allocations for instance 0da17f71-46ee-42f3-99f7-62f4db6e1d9d [ 689.994703] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dbfe3ba-3127-4880-a62f-42546750dfba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.003202] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0826701b-191c-475c-8709-52185871fd78 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.034980] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a16d4e-80d1-4f40-92fc-cb1fc1f05c40 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.042501] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da586b26-c6cb-4673-88a4-892ded782f04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.055903] env[61907]: DEBUG nova.compute.provider_tree [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.107613] env[61907]: DEBUG nova.network.neutron [-] [instance: a1189e41-5922-4623-bb03-394d98159193] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.329837] env[61907]: DEBUG nova.network.neutron [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.484966] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e2c8c750-8134-448c-b510-c6215b085957 tempest-FloatingIPsAssociationNegativeTestJSON-712046511 tempest-FloatingIPsAssociationNegativeTestJSON-712046511-project-member] Lock "0da17f71-46ee-42f3-99f7-62f4db6e1d9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.089s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.561991] env[61907]: DEBUG nova.scheduler.client.report [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 690.609723] env[61907]: INFO nova.compute.manager [-] [instance: a1189e41-5922-4623-bb03-394d98159193] Took 1.02 seconds to deallocate network for instance. [ 690.612194] env[61907]: DEBUG nova.compute.claims [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 690.612372] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.832162] env[61907]: INFO nova.compute.manager [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: ab7d4c93-fbc9-4446-ac57-f479ab408180] Took 1.02 seconds to deallocate network for instance. [ 690.990536] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 691.067189] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.067828] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 691.070749] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.165s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.072682] env[61907]: INFO nova.compute.claims [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.519630] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.577271] env[61907]: DEBUG nova.compute.utils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.578756] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 691.578929] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.626437] env[61907]: DEBUG nova.policy [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c800100155594a18b23b0ccc4b25b62d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1b55e8d6e07740d480c81a08aef84233', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.866815] env[61907]: INFO nova.scheduler.client.report [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocations for instance ab7d4c93-fbc9-4446-ac57-f479ab408180 [ 691.946777] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Successfully created port: 9f6e0504-22f2-4094-b573-fdb5164cf4a1 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.085281] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 692.379771] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e76541ec-4165-4f0d-90d7-a03016f82f26 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "ab7d4c93-fbc9-4446-ac57-f479ab408180" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.695s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.468158] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7baf9b8-a194-4496-a5ed-9b5f37bedf89 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.476861] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054c89b0-bb86-4813-b217-730c1149d948 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.512061] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d63296a-147d-4d83-97b9-166808b50abe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.516818] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98f98cd1-ec3c-4bc7-9eac-deda7402e07e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.530710] env[61907]: DEBUG nova.compute.provider_tree [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.881257] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 693.022355] env[61907]: DEBUG nova.compute.manager [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Received event network-changed-9f6e0504-22f2-4094-b573-fdb5164cf4a1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 693.022553] env[61907]: DEBUG nova.compute.manager [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Refreshing instance network info cache due to event network-changed-9f6e0504-22f2-4094-b573-fdb5164cf4a1. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 693.022765] env[61907]: DEBUG oslo_concurrency.lockutils [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] Acquiring lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.022924] env[61907]: DEBUG oslo_concurrency.lockutils [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] Acquired lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.023116] env[61907]: DEBUG nova.network.neutron [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Refreshing network info cache for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.033421] env[61907]: DEBUG nova.scheduler.client.report [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 693.107847] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 693.135655] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.135792] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.135890] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.136083] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.136682] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.136682] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.136682] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.136823] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.136926] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.137102] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.137328] env[61907]: DEBUG nova.virt.hardware [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.138238] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0db4751-37e2-4a15-bea1-f8524266f3e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.146149] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c90738a1-d2c2-48d1-81a1-a7f453917f08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.186496] env[61907]: ERROR nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 693.186496] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 693.186496] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.186496] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.186496] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.186496] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.186496] env[61907]: ERROR nova.compute.manager raise self.value [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.186496] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.186496] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.186496] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.186971] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.186971] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.186971] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 693.186971] env[61907]: ERROR nova.compute.manager [ 693.186971] env[61907]: Traceback (most recent call last): [ 693.186971] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.186971] env[61907]: listener.cb(fileno) [ 693.186971] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.186971] env[61907]: result = function(*args, **kwargs) [ 693.186971] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.186971] env[61907]: return func(*args, **kwargs) [ 693.186971] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 693.186971] env[61907]: raise e [ 693.186971] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 693.186971] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 693.186971] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.186971] env[61907]: created_port_ids = self._update_ports_for_instance( [ 693.186971] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.186971] env[61907]: with excutils.save_and_reraise_exception(): [ 693.186971] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.186971] env[61907]: self.force_reraise() [ 693.186971] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.186971] env[61907]: raise self.value [ 693.186971] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.186971] env[61907]: updated_port = self._update_port( [ 693.186971] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.186971] env[61907]: _ensure_no_port_binding_failure(port) [ 693.186971] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.186971] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.188530] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 693.188530] env[61907]: Removing descriptor: 17 [ 693.188530] env[61907]: ERROR nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Traceback (most recent call last): [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] yield resources [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.driver.spawn(context, instance, image_meta, [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.188530] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] vm_ref = self.build_virtual_machine(instance, [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] for vif in network_info: [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self._sync_wrapper(fn, *args, **kwargs) [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.wait() [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self[:] = self._gt.wait() [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self._exit_event.wait() [ 693.188840] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] result = hub.switch() [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self.greenlet.switch() [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] result = function(*args, **kwargs) [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return func(*args, **kwargs) [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise e [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] nwinfo = self.network_api.allocate_for_instance( [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.189609] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] created_port_ids = self._update_ports_for_instance( [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] with excutils.save_and_reraise_exception(): [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.force_reraise() [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise self.value [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] updated_port = self._update_port( [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] _ensure_no_port_binding_failure(port) [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.189965] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise exception.PortBindingFailed(port_id=port['id']) [ 693.190387] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 693.190387] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] [ 693.190387] env[61907]: INFO nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Terminating instance [ 693.415436] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.538262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.539018] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 693.542200] env[61907]: DEBUG nova.network.neutron [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.543912] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.319s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.650836] env[61907]: DEBUG nova.network.neutron [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.692765] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquiring lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.049603] env[61907]: DEBUG nova.compute.utils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 694.054058] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 694.054882] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 694.092600] env[61907]: DEBUG nova.policy [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c2f7b4952192462baaf1b82a70228ab4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ea36441e28a4f7ab32598884100197f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 694.122054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.122159] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.154379] env[61907]: DEBUG oslo_concurrency.lockutils [req-5dc835a8-1682-4231-b596-549a711ddf06 req-9964f86b-081c-4436-9803-0c35be7137e3 service nova] Releasing lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.154749] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquired lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.154928] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.369471] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc522568-391a-4751-8e4c-47c06283e6b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.377213] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Successfully created port: b3690168-eed1-45bf-90fb-d078abe58053 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 694.379653] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a1a42c-54f8-4ad7-b819-55b48e84cdb1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.413488] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eaa5f47-e73c-4d3f-8f17-9cebaca4b737 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.421569] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9fe709c-4ad8-4d8a-9178-74ff42c1f907 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.435908] env[61907]: DEBUG nova.compute.provider_tree [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.555176] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 694.684520] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.819707] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.938641] env[61907]: DEBUG nova.scheduler.client.report [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 695.069717] env[61907]: DEBUG nova.compute.manager [req-925ba11a-3613-446d-b76f-7b7a6a46cb39 req-b32eb1a0-eac1-4a43-8cda-528505aec2d7 service nova] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Received event network-vif-deleted-9f6e0504-22f2-4094-b573-fdb5164cf4a1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 695.322570] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Releasing lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 695.323078] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 695.323288] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 695.323656] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41ba4f0f-ca8b-42a7-98db-cca6fe5478d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.332819] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-547eb44e-afab-4bf8-a557-4aaa75ebb79f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.358020] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee could not be found. [ 695.358020] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.358020] env[61907]: INFO nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Took 0.03 seconds to destroy the instance on the hypervisor. [ 695.358020] env[61907]: DEBUG oslo.service.loopingcall [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.358020] env[61907]: DEBUG nova.compute.manager [-] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 695.358020] env[61907]: DEBUG nova.network.neutron [-] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.443876] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.900s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.444545] env[61907]: ERROR nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Traceback (most recent call last): [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.driver.spawn(context, instance, image_meta, [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] vm_ref = self.build_virtual_machine(instance, [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.444545] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] for vif in network_info: [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return self._sync_wrapper(fn, *args, **kwargs) [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.wait() [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self[:] = self._gt.wait() [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return self._exit_event.wait() [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] current.throw(*self._exc) [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.445019] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] result = function(*args, **kwargs) [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] return func(*args, **kwargs) [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise e [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] nwinfo = self.network_api.allocate_for_instance( [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] created_port_ids = self._update_ports_for_instance( [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] with excutils.save_and_reraise_exception(): [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] self.force_reraise() [ 695.445530] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise self.value [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] updated_port = self._update_port( [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] _ensure_no_port_binding_failure(port) [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] raise exception.PortBindingFailed(port_id=port['id']) [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] nova.exception.PortBindingFailed: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. [ 695.446036] env[61907]: ERROR nova.compute.manager [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] [ 695.446036] env[61907]: DEBUG nova.compute.utils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.446481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.863s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.449344] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Build of instance 78f3d457-d76b-4a13-9f12-264e53b768b4 was re-scheduled: Binding failed for port 87d5456d-e601-4bbd-bba5-66683873c2f3, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 695.449757] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 695.449977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquiring lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.450187] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Acquired lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.450339] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 695.504222] env[61907]: DEBUG nova.network.neutron [-] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.539612] env[61907]: ERROR nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 695.539612] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.539612] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.539612] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.539612] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.539612] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.539612] env[61907]: ERROR nova.compute.manager raise self.value [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.539612] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 695.539612] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.539612] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 695.540257] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.540257] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 695.540257] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 695.540257] env[61907]: ERROR nova.compute.manager [ 695.540257] env[61907]: Traceback (most recent call last): [ 695.540257] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 695.540257] env[61907]: listener.cb(fileno) [ 695.540257] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.540257] env[61907]: result = function(*args, **kwargs) [ 695.540257] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.540257] env[61907]: return func(*args, **kwargs) [ 695.540257] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.540257] env[61907]: raise e [ 695.540257] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.540257] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 695.540257] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.540257] env[61907]: created_port_ids = self._update_ports_for_instance( [ 695.540257] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.540257] env[61907]: with excutils.save_and_reraise_exception(): [ 695.540257] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.540257] env[61907]: self.force_reraise() [ 695.540257] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.540257] env[61907]: raise self.value [ 695.540257] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.540257] env[61907]: updated_port = self._update_port( [ 695.540257] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.540257] env[61907]: _ensure_no_port_binding_failure(port) [ 695.540257] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.540257] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 695.541035] env[61907]: nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 695.541035] env[61907]: Removing descriptor: 17 [ 695.563875] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 695.589359] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 695.589639] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 695.589798] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 695.589981] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 695.590153] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 695.590321] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 695.590539] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 695.590701] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 695.590866] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 695.591046] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 695.591231] env[61907]: DEBUG nova.virt.hardware [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 695.592120] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db45fafe-ef72-4d68-88be-f1cc6553dcd7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.601115] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0620fe3e-bc9b-4019-8e87-8cbf54e61f41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.613817] env[61907]: ERROR nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Traceback (most recent call last): [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] yield resources [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.driver.spawn(context, instance, image_meta, [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] vm_ref = self.build_virtual_machine(instance, [ 695.613817] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] for vif in network_info: [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return self._sync_wrapper(fn, *args, **kwargs) [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.wait() [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self[:] = self._gt.wait() [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return self._exit_event.wait() [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 695.614345] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] current.throw(*self._exc) [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] result = function(*args, **kwargs) [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return func(*args, **kwargs) [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise e [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] nwinfo = self.network_api.allocate_for_instance( [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] created_port_ids = self._update_ports_for_instance( [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] with excutils.save_and_reraise_exception(): [ 695.614872] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.force_reraise() [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise self.value [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] updated_port = self._update_port( [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] _ensure_no_port_binding_failure(port) [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise exception.PortBindingFailed(port_id=port['id']) [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 695.615309] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] [ 695.615309] env[61907]: INFO nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Terminating instance [ 695.973234] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.006495] env[61907]: DEBUG nova.network.neutron [-] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.069114] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.120075] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquiring lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.120375] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquired lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.120477] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 696.269779] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94253ed1-9b4a-4beb-95a9-bd2afadf7344 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.277379] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c53644a-44f7-4289-8fe3-9dcfe664cc91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.307167] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9d01f3-7ed2-4860-ad92-d6539ffeec22 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.313608] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe513efa-69b2-44bf-af23-c411ca332c8b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.326178] env[61907]: DEBUG nova.compute.provider_tree [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 696.510634] env[61907]: INFO nova.compute.manager [-] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Took 1.15 seconds to deallocate network for instance. [ 696.513011] env[61907]: DEBUG nova.compute.claims [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.513204] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.574086] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Releasing lock "refresh_cache-78f3d457-d76b-4a13-9f12-264e53b768b4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.574397] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 696.574577] env[61907]: DEBUG nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 696.574744] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 696.598676] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.646703] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 696.735065] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.830123] env[61907]: DEBUG nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 697.101865] env[61907]: DEBUG nova.network.neutron [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.112746] env[61907]: DEBUG nova.compute.manager [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Received event network-changed-b3690168-eed1-45bf-90fb-d078abe58053 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 697.112927] env[61907]: DEBUG nova.compute.manager [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Refreshing instance network info cache due to event network-changed-b3690168-eed1-45bf-90fb-d078abe58053. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 697.113116] env[61907]: DEBUG oslo_concurrency.lockutils [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] Acquiring lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.237571] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Releasing lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.237844] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 697.237961] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 697.238301] env[61907]: DEBUG oslo_concurrency.lockutils [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] Acquired lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.238478] env[61907]: DEBUG nova.network.neutron [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Refreshing network info cache for port b3690168-eed1-45bf-90fb-d078abe58053 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 697.239584] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-99888b8e-ba75-458b-8b08-7f0ef9f57d26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.249424] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bc054e-bab7-47ee-9785-dc867e973c58 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.271838] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d could not be found. [ 697.272122] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 697.272317] env[61907]: INFO nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Took 0.03 seconds to destroy the instance on the hypervisor. [ 697.272553] env[61907]: DEBUG oslo.service.loopingcall [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.272855] env[61907]: DEBUG nova.compute.manager [-] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 697.272855] env[61907]: DEBUG nova.network.neutron [-] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 697.287484] env[61907]: DEBUG nova.network.neutron [-] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.333566] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.334191] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Traceback (most recent call last): [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.driver.spawn(context, instance, image_meta, [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self._vmops.spawn(context, instance, image_meta, injected_files, [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] vm_ref = self.build_virtual_machine(instance, [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] vif_infos = vmwarevif.get_vif_info(self._session, [ 697.334191] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] for vif in network_info: [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return self._sync_wrapper(fn, *args, **kwargs) [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.wait() [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self[:] = self._gt.wait() [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return self._exit_event.wait() [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] current.throw(*self._exc) [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 697.334528] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] result = function(*args, **kwargs) [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] return func(*args, **kwargs) [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise e [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] nwinfo = self.network_api.allocate_for_instance( [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] created_port_ids = self._update_ports_for_instance( [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] with excutils.save_and_reraise_exception(): [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] self.force_reraise() [ 697.334906] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise self.value [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] updated_port = self._update_port( [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] _ensure_no_port_binding_failure(port) [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] raise exception.PortBindingFailed(port_id=port['id']) [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] nova.exception.PortBindingFailed: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. [ 697.335286] env[61907]: ERROR nova.compute.manager [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] [ 697.335286] env[61907]: DEBUG nova.compute.utils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.336327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.837s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.339047] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Build of instance 90a4f8bb-59bd-499f-b2fc-ebce55c5f803 was re-scheduled: Binding failed for port 0f90d605-9722-4138-9823-c17442559052, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 697.339494] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 697.339713] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.339856] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.340015] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 697.604511] env[61907]: INFO nova.compute.manager [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] [instance: 78f3d457-d76b-4a13-9f12-264e53b768b4] Took 1.03 seconds to deallocate network for instance. [ 697.758467] env[61907]: DEBUG nova.network.neutron [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.790332] env[61907]: DEBUG nova.network.neutron [-] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.847412] env[61907]: DEBUG nova.network.neutron [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.859081] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 697.954310] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.156329] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e6a9ca-b4fb-4961-971d-dc7bf3737075 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.164077] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb0f417f-1fe2-44bd-9253-b525353f70af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.194598] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9093371-d6dd-4f7c-9beb-c7fd0b8edd61 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.202281] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-720e1e03-3cd4-45ca-8718-fa516d50062a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.215695] env[61907]: DEBUG nova.compute.provider_tree [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 698.293328] env[61907]: INFO nova.compute.manager [-] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Took 1.02 seconds to deallocate network for instance. [ 698.295479] env[61907]: DEBUG nova.compute.claims [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 698.295655] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.350358] env[61907]: DEBUG oslo_concurrency.lockutils [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] Releasing lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.350565] env[61907]: DEBUG nova.compute.manager [req-31d61275-3f2a-44bc-a44e-a242748e6466 req-d61a89ec-55a7-4efd-a285-b870820f05ea service nova] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Received event network-vif-deleted-b3690168-eed1-45bf-90fb-d078abe58053 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 698.457482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-90a4f8bb-59bd-499f-b2fc-ebce55c5f803" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.457725] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 698.457907] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 698.458090] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 698.474019] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.634010] env[61907]: INFO nova.scheduler.client.report [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Deleted allocations for instance 78f3d457-d76b-4a13-9f12-264e53b768b4 [ 698.720847] env[61907]: DEBUG nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 698.976904] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.142484] env[61907]: DEBUG oslo_concurrency.lockutils [None req-935fd673-d038-48ca-a252-c99b80aebb1c tempest-ServerActionsV293TestJSON-145349524 tempest-ServerActionsV293TestJSON-145349524-project-member] Lock "78f3d457-d76b-4a13-9f12-264e53b768b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.999s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.226433] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.227247] env[61907]: ERROR nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Traceback (most recent call last): [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.driver.spawn(context, instance, image_meta, [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] vm_ref = self.build_virtual_machine(instance, [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.227247] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] for vif in network_info: [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return self._sync_wrapper(fn, *args, **kwargs) [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.wait() [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self[:] = self._gt.wait() [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return self._exit_event.wait() [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] current.throw(*self._exc) [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.227701] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] result = function(*args, **kwargs) [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] return func(*args, **kwargs) [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise e [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] nwinfo = self.network_api.allocate_for_instance( [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] created_port_ids = self._update_ports_for_instance( [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] with excutils.save_and_reraise_exception(): [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] self.force_reraise() [ 699.228031] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise self.value [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] updated_port = self._update_port( [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] _ensure_no_port_binding_failure(port) [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] raise exception.PortBindingFailed(port_id=port['id']) [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] nova.exception.PortBindingFailed: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. [ 699.228361] env[61907]: ERROR nova.compute.manager [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] [ 699.228361] env[61907]: DEBUG nova.compute.utils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.229790] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.498s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.233269] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Build of instance 6ac3cdf2-eb93-4403-a514-81d56baf544e was re-scheduled: Binding failed for port 0164397e-bb5a-4c6e-8730-ec31fbda35ba, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 699.233901] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 699.234208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.234448] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.234701] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.479960] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 90a4f8bb-59bd-499f-b2fc-ebce55c5f803] Took 1.02 seconds to deallocate network for instance. [ 699.647534] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 699.785200] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 699.871939] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.090112] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa39b688-ba4d-4e3b-93fe-58c12562b5ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.098065] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8daa2c2-686f-4898-b171-659f4e90b531 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.129477] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9e1351-6037-457f-bbaa-7f63c85010df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.136429] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e3d8c76-6703-4e0c-8a8f-2dcb5801dcba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.153966] env[61907]: DEBUG nova.compute.provider_tree [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.171310] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.374327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-6ac3cdf2-eb93-4403-a514-81d56baf544e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.374590] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 700.374770] env[61907]: DEBUG nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 700.374941] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 700.390556] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.511775] env[61907]: INFO nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted allocations for instance 90a4f8bb-59bd-499f-b2fc-ebce55c5f803 [ 700.661025] env[61907]: DEBUG nova.scheduler.client.report [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 700.893912] env[61907]: DEBUG nova.network.neutron [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.019678] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "90a4f8bb-59bd-499f-b2fc-ebce55c5f803" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.519s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.166401] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.937s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.166816] env[61907]: ERROR nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Traceback (most recent call last): [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.driver.spawn(context, instance, image_meta, [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] vm_ref = self.build_virtual_machine(instance, [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] vif_infos = vmwarevif.get_vif_info(self._session, [ 701.166816] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] for vif in network_info: [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return self._sync_wrapper(fn, *args, **kwargs) [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.wait() [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self[:] = self._gt.wait() [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return self._exit_event.wait() [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] current.throw(*self._exc) [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 701.167137] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] result = function(*args, **kwargs) [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] return func(*args, **kwargs) [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise e [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] nwinfo = self.network_api.allocate_for_instance( [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] created_port_ids = self._update_ports_for_instance( [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] with excutils.save_and_reraise_exception(): [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] self.force_reraise() [ 701.167787] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise self.value [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] updated_port = self._update_port( [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] _ensure_no_port_binding_failure(port) [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] raise exception.PortBindingFailed(port_id=port['id']) [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] nova.exception.PortBindingFailed: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. [ 701.168149] env[61907]: ERROR nova.compute.manager [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] [ 701.168149] env[61907]: DEBUG nova.compute.utils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 701.169179] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Build of instance 8456ccc3-aa32-4388-bf9e-2608c783405f was re-scheduled: Binding failed for port c1ba743f-2581-4733-9fb3-5f239b4b4609, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 701.169716] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 701.169996] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.170097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.170264] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 701.171238] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.971s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.397402] env[61907]: INFO nova.compute.manager [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 6ac3cdf2-eb93-4403-a514-81d56baf544e] Took 1.02 seconds to deallocate network for instance. [ 701.525861] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 701.700615] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.793284] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.959348] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0650dbe5-f137-4374-8d90-67833f9f85fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.966905] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647730fe-952f-4731-b22c-afcec1236b64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.996202] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cccc39f-e51d-47eb-9953-d069c5d3378c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.002740] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87291007-f491-4840-887d-2a61f8859c85 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.016838] env[61907]: DEBUG nova.compute.provider_tree [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.041503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.295549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-8456ccc3-aa32-4388-bf9e-2608c783405f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.295803] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 702.296009] env[61907]: DEBUG nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 702.296192] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 702.311793] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.434790] env[61907]: INFO nova.scheduler.client.report [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted allocations for instance 6ac3cdf2-eb93-4403-a514-81d56baf544e [ 702.520025] env[61907]: DEBUG nova.scheduler.client.report [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 702.813990] env[61907]: DEBUG nova.network.neutron [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.943280] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27b6225c-f0c1-413c-81ff-6f2a3cb7c23c tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "6ac3cdf2-eb93-4403-a514-81d56baf544e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.400s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.024707] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.853s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.025254] env[61907]: ERROR nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Traceback (most recent call last): [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.driver.spawn(context, instance, image_meta, [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] vm_ref = self.build_virtual_machine(instance, [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.025254] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] for vif in network_info: [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return self._sync_wrapper(fn, *args, **kwargs) [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.wait() [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self[:] = self._gt.wait() [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return self._exit_event.wait() [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] current.throw(*self._exc) [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.025606] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] result = function(*args, **kwargs) [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] return func(*args, **kwargs) [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise e [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] nwinfo = self.network_api.allocate_for_instance( [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] created_port_ids = self._update_ports_for_instance( [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] with excutils.save_and_reraise_exception(): [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] self.force_reraise() [ 703.025980] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise self.value [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] updated_port = self._update_port( [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] _ensure_no_port_binding_failure(port) [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] raise exception.PortBindingFailed(port_id=port['id']) [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] nova.exception.PortBindingFailed: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. [ 703.026362] env[61907]: ERROR nova.compute.manager [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] [ 703.026362] env[61907]: DEBUG nova.compute.utils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.027515] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.415s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.031239] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Build of instance 520e2134-3553-44c2-98f5-ecd5be642d86 was re-scheduled: Binding failed for port c87a7e7d-7c2f-41c5-bcf5-e0951999a913, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 703.031386] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 703.032415] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.032415] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.032415] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.317163] env[61907]: INFO nova.compute.manager [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 8456ccc3-aa32-4388-bf9e-2608c783405f] Took 1.02 seconds to deallocate network for instance. [ 703.445618] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 703.559392] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.662158] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.827293] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5145a285-3756-49ad-852d-0e9d16b4c7b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.836352] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45c3ac69-0059-4cdf-9776-71f1cf2fdce0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.867535] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd7e62b-d326-46d1-8f47-81dcd0e64a0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.875380] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddae77cb-7d1e-40f1-9855-09075a5e1add {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.890181] env[61907]: DEBUG nova.compute.provider_tree [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 703.965050] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.166799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-520e2134-3553-44c2-98f5-ecd5be642d86" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.167125] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 704.167255] env[61907]: DEBUG nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 704.167470] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.182444] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.344294] env[61907]: INFO nova.scheduler.client.report [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted allocations for instance 8456ccc3-aa32-4388-bf9e-2608c783405f [ 704.395625] env[61907]: DEBUG nova.scheduler.client.report [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 704.685261] env[61907]: DEBUG nova.network.neutron [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.734123] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.734352] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.760036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "08647e48-9e24-4d7c-a9fd-4066317a7250" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.760269] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.853988] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b0e3070-0a4d-4edb-a8a4-fcb3684539f2 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "8456ccc3-aa32-4388-bf9e-2608c783405f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 106.094s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.898407] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.871s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.899044] env[61907]: ERROR nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] Traceback (most recent call last): [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.driver.spawn(context, instance, image_meta, [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self._vmops.spawn(context, instance, image_meta, injected_files, [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] vm_ref = self.build_virtual_machine(instance, [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] vif_infos = vmwarevif.get_vif_info(self._session, [ 704.899044] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] for vif in network_info: [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self._sync_wrapper(fn, *args, **kwargs) [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.wait() [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self[:] = self._gt.wait() [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self._exit_event.wait() [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] result = hub.switch() [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 704.899406] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return self.greenlet.switch() [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] result = function(*args, **kwargs) [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] return func(*args, **kwargs) [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise e [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] nwinfo = self.network_api.allocate_for_instance( [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] created_port_ids = self._update_ports_for_instance( [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] with excutils.save_and_reraise_exception(): [ 704.899722] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] self.force_reraise() [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise self.value [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] updated_port = self._update_port( [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] _ensure_no_port_binding_failure(port) [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] raise exception.PortBindingFailed(port_id=port['id']) [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] nova.exception.PortBindingFailed: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. [ 704.900064] env[61907]: ERROR nova.compute.manager [instance: a1189e41-5922-4623-bb03-394d98159193] [ 704.900336] env[61907]: DEBUG nova.compute.utils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 704.900985] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.381s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.902445] env[61907]: INFO nova.compute.claims [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 704.905037] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Build of instance a1189e41-5922-4623-bb03-394d98159193 was re-scheduled: Binding failed for port cba547b3-4808-4176-bf9e-a581c2c6c6c0, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 704.905575] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 704.905650] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquiring lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.905761] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Acquired lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.905987] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 705.187645] env[61907]: INFO nova.compute.manager [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 520e2134-3553-44c2-98f5-ecd5be642d86] Took 1.02 seconds to deallocate network for instance. [ 705.356402] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 705.427829] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 705.521324] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.879245] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.023604] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Releasing lock "refresh_cache-a1189e41-5922-4623-bb03-394d98159193" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.023851] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 706.024048] env[61907]: DEBUG nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 706.024222] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 706.046918] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 706.185404] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2167496-fa0a-4b3f-aed6-88ee0549478d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.193652] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25290417-3eed-4211-89f3-fb564dc5891f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.228471] env[61907]: INFO nova.scheduler.client.report [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted allocations for instance 520e2134-3553-44c2-98f5-ecd5be642d86 [ 706.235153] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7742cc6f-cd55-4133-bc3d-66b99f4f9415 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.245205] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27553e6e-541d-4f94-b2d4-9afdad623a25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.263453] env[61907]: DEBUG nova.compute.provider_tree [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.515263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "ce864af6-aef6-4044-be64-8440ba175438" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 706.515498] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.549578] env[61907]: DEBUG nova.network.neutron [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.740804] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27e32142-4052-4991-9acd-2a66a0d71188 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "520e2134-3553-44c2-98f5-ecd5be642d86" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.686s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.766572] env[61907]: DEBUG nova.scheduler.client.report [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 707.053143] env[61907]: INFO nova.compute.manager [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] [instance: a1189e41-5922-4623-bb03-394d98159193] Took 1.03 seconds to deallocate network for instance. [ 707.248237] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 707.276628] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.371s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.276628] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 707.276628] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.860s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.277140] env[61907]: INFO nova.compute.claims [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.772836] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.781357] env[61907]: DEBUG nova.compute.utils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.785042] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 707.785042] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 707.846987] env[61907]: DEBUG nova.policy [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b73c580c9cf44ebac33a16e7e098d96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f52cec3f47b84d78ae6cf9bc881784b3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.085191] env[61907]: INFO nova.scheduler.client.report [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Deleted allocations for instance a1189e41-5922-4623-bb03-394d98159193 [ 708.225020] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Successfully created port: c801028f-2d97-4011-bdf5-cd27de91e138 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.284066] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 708.596619] env[61907]: DEBUG oslo_concurrency.lockutils [None req-37956720-405e-42bd-a0e7-55019696b97e tempest-ServersTestFqdnHostnames-1405257273 tempest-ServersTestFqdnHostnames-1405257273-project-member] Lock "a1189e41-5922-4623-bb03-394d98159193" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.273s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.651885] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03713bef-19eb-45a3-9392-58d34012af44 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.659442] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bf47ad-c21b-4cc7-b858-44f6be245fbc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.692018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387e1c8d-b5ea-4e7b-9419-4f8678c6a22d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.698621] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c820640-7558-4b70-b362-5614f110d936 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.711853] env[61907]: DEBUG nova.compute.provider_tree [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.099089] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 709.164844] env[61907]: DEBUG nova.compute.manager [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Received event network-changed-c801028f-2d97-4011-bdf5-cd27de91e138 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 709.165047] env[61907]: DEBUG nova.compute.manager [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Refreshing instance network info cache due to event network-changed-c801028f-2d97-4011-bdf5-cd27de91e138. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 709.165282] env[61907]: DEBUG oslo_concurrency.lockutils [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] Acquiring lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.165424] env[61907]: DEBUG oslo_concurrency.lockutils [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] Acquired lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.165607] env[61907]: DEBUG nova.network.neutron [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Refreshing network info cache for port c801028f-2d97-4011-bdf5-cd27de91e138 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 709.215541] env[61907]: DEBUG nova.scheduler.client.report [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 709.292819] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 709.319367] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.319686] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.319851] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.320048] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.320199] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.320347] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.320557] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.320718] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.320884] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.321230] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.321473] env[61907]: DEBUG nova.virt.hardware [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.322540] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ad56a4-44d0-4c49-b16d-87606abab69b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.331493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918688c3-c032-4fd2-b9a8-c27f3bd95b61 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.419301] env[61907]: ERROR nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 709.419301] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 709.419301] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.419301] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.419301] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.419301] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.419301] env[61907]: ERROR nova.compute.manager raise self.value [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.419301] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 709.419301] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.419301] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 709.419810] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.419810] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 709.419810] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 709.419810] env[61907]: ERROR nova.compute.manager [ 709.419810] env[61907]: Traceback (most recent call last): [ 709.419810] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 709.419810] env[61907]: listener.cb(fileno) [ 709.419810] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.419810] env[61907]: result = function(*args, **kwargs) [ 709.419810] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.419810] env[61907]: return func(*args, **kwargs) [ 709.419810] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 709.419810] env[61907]: raise e [ 709.419810] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 709.419810] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 709.419810] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.419810] env[61907]: created_port_ids = self._update_ports_for_instance( [ 709.419810] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.419810] env[61907]: with excutils.save_and_reraise_exception(): [ 709.419810] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.419810] env[61907]: self.force_reraise() [ 709.419810] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.419810] env[61907]: raise self.value [ 709.419810] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.419810] env[61907]: updated_port = self._update_port( [ 709.419810] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.419810] env[61907]: _ensure_no_port_binding_failure(port) [ 709.419810] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.419810] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 709.420683] env[61907]: nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 709.420683] env[61907]: Removing descriptor: 15 [ 709.420683] env[61907]: ERROR nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Traceback (most recent call last): [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] yield resources [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.driver.spawn(context, instance, image_meta, [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.420683] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] vm_ref = self.build_virtual_machine(instance, [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] for vif in network_info: [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self._sync_wrapper(fn, *args, **kwargs) [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.wait() [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self[:] = self._gt.wait() [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self._exit_event.wait() [ 709.421064] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] result = hub.switch() [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self.greenlet.switch() [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] result = function(*args, **kwargs) [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return func(*args, **kwargs) [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise e [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] nwinfo = self.network_api.allocate_for_instance( [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.421448] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] created_port_ids = self._update_ports_for_instance( [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] with excutils.save_and_reraise_exception(): [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.force_reraise() [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise self.value [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] updated_port = self._update_port( [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] _ensure_no_port_binding_failure(port) [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.421846] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise exception.PortBindingFailed(port_id=port['id']) [ 709.423265] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 709.423265] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] [ 709.423265] env[61907]: INFO nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Terminating instance [ 709.623372] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.684211] env[61907]: DEBUG nova.network.neutron [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.721290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.446s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.721906] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 709.726549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.213s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.766223] env[61907]: DEBUG nova.network.neutron [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.924357] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquiring lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.018735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.018735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.231980] env[61907]: DEBUG nova.compute.utils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.239341] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 710.239645] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.271030] env[61907]: DEBUG oslo_concurrency.lockutils [req-81afc326-9484-4eb0-a04d-d51cea92e994 req-39363993-75ab-49d2-bde1-a7f2afa5c64c service nova] Releasing lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.271311] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquired lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.271454] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.308279] env[61907]: DEBUG nova.policy [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0745691ba6c44a42b887c649d7752ebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c808b1baf1842fbb9e2d28f0031e4d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.638023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efaee0b0-52b5-4b31-be44-079cdb9780a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.644711] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b22e31b5-8b8c-47ea-bf31-80058f983051 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.681935] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c7b789-2c8d-44dd-851f-9cb693938375 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.694201] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c34de6-c5d8-4721-97b2-f59ce0b3fc80 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.710655] env[61907]: DEBUG nova.compute.provider_tree [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.738949] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 710.798324] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.974266] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.974395] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Successfully created port: a534c8b5-7d31-46ad-accf-f786637d2194 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.188275] env[61907]: DEBUG nova.compute.manager [req-b70ecafd-432e-4db1-be51-787dd7468bf4 req-d8c82969-8132-4d8c-ab3d-66357e3189ca service nova] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Received event network-vif-deleted-c801028f-2d97-4011-bdf5-cd27de91e138 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 711.218562] env[61907]: DEBUG nova.scheduler.client.report [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 711.479549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Releasing lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.479806] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 711.482622] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.482852] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f82fd612-81e6-44f4-854c-99b2b199effd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.494982] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32044d91-94ef-47d2-9825-618677b18944 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.520895] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5 could not be found. [ 711.521149] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.521338] env[61907]: INFO nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 711.521582] env[61907]: DEBUG oslo.service.loopingcall [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.521931] env[61907]: DEBUG nova.compute.manager [-] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 711.521931] env[61907]: DEBUG nova.network.neutron [-] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.543314] env[61907]: DEBUG nova.network.neutron [-] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.724455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.725717] env[61907]: ERROR nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Traceback (most recent call last): [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.driver.spawn(context, instance, image_meta, [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] vm_ref = self.build_virtual_machine(instance, [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.725717] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] for vif in network_info: [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self._sync_wrapper(fn, *args, **kwargs) [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.wait() [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self[:] = self._gt.wait() [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self._exit_event.wait() [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] result = hub.switch() [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.726035] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return self.greenlet.switch() [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] result = function(*args, **kwargs) [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] return func(*args, **kwargs) [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise e [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] nwinfo = self.network_api.allocate_for_instance( [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] created_port_ids = self._update_ports_for_instance( [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] with excutils.save_and_reraise_exception(): [ 711.726367] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] self.force_reraise() [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise self.value [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] updated_port = self._update_port( [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] _ensure_no_port_binding_failure(port) [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] raise exception.PortBindingFailed(port_id=port['id']) [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] nova.exception.PortBindingFailed: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. [ 711.726794] env[61907]: ERROR nova.compute.manager [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] [ 711.727155] env[61907]: DEBUG nova.compute.utils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.727930] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.432s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.731046] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Build of instance a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee was re-scheduled: Binding failed for port 9f6e0504-22f2-4094-b573-fdb5164cf4a1, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 711.731488] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 711.731713] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquiring lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.731860] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Acquired lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.732025] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.749721] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 711.779646] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.779910] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.780081] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.780268] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.780424] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.780593] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.780804] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.780963] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.781302] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.781484] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.781662] env[61907]: DEBUG nova.virt.hardware [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.782748] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5f6cea-2d33-4361-a3f2-4166ecdd59a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.790550] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9aa842-589e-4682-8bff-71df2b58bdd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.028241] env[61907]: ERROR nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 712.028241] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.028241] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.028241] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.028241] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.028241] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.028241] env[61907]: ERROR nova.compute.manager raise self.value [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.028241] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 712.028241] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.028241] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 712.028676] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.028676] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 712.028676] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 712.028676] env[61907]: ERROR nova.compute.manager [ 712.028676] env[61907]: Traceback (most recent call last): [ 712.028676] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 712.028676] env[61907]: listener.cb(fileno) [ 712.028676] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.028676] env[61907]: result = function(*args, **kwargs) [ 712.028676] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.028676] env[61907]: return func(*args, **kwargs) [ 712.028676] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 712.028676] env[61907]: raise e [ 712.028676] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.028676] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 712.028676] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.028676] env[61907]: created_port_ids = self._update_ports_for_instance( [ 712.028676] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.028676] env[61907]: with excutils.save_and_reraise_exception(): [ 712.028676] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.028676] env[61907]: self.force_reraise() [ 712.028676] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.028676] env[61907]: raise self.value [ 712.028676] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.028676] env[61907]: updated_port = self._update_port( [ 712.028676] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.028676] env[61907]: _ensure_no_port_binding_failure(port) [ 712.028676] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.028676] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 712.029531] env[61907]: nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 712.029531] env[61907]: Removing descriptor: 15 [ 712.029531] env[61907]: ERROR nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Traceback (most recent call last): [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] yield resources [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.driver.spawn(context, instance, image_meta, [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.029531] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] vm_ref = self.build_virtual_machine(instance, [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] for vif in network_info: [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self._sync_wrapper(fn, *args, **kwargs) [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.wait() [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self[:] = self._gt.wait() [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self._exit_event.wait() [ 712.029904] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] result = hub.switch() [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self.greenlet.switch() [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] result = function(*args, **kwargs) [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return func(*args, **kwargs) [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise e [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] nwinfo = self.network_api.allocate_for_instance( [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.030281] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] created_port_ids = self._update_ports_for_instance( [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] with excutils.save_and_reraise_exception(): [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.force_reraise() [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise self.value [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] updated_port = self._update_port( [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] _ensure_no_port_binding_failure(port) [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.030648] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise exception.PortBindingFailed(port_id=port['id']) [ 712.031015] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 712.031015] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] [ 712.031015] env[61907]: INFO nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Terminating instance [ 712.045769] env[61907]: DEBUG nova.network.neutron [-] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.263099] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.363388] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.538954] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.539262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.539394] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.548715] env[61907]: INFO nova.compute.manager [-] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Took 1.03 seconds to deallocate network for instance. [ 712.553713] env[61907]: DEBUG nova.compute.claims [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.553895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.564042] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319e4af2-0b7a-42fd-a7ba-014642f84c6f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.574681] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de9139a6-3401-49ed-b761-263b6287cc35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.605534] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1098dd20-3f65-4cbd-9aad-854ac49f4b64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.613036] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a960e8fc-0ac1-4aa5-825f-749fffdfcd90 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.626495] env[61907]: DEBUG nova.compute.provider_tree [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.866324] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Releasing lock "refresh_cache-a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.866592] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 712.866747] env[61907]: DEBUG nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 712.866917] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.882369] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.076817] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.129537] env[61907]: DEBUG nova.scheduler.client.report [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 713.177583] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.293512] env[61907]: DEBUG nova.compute.manager [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Received event network-changed-a534c8b5-7d31-46ad-accf-f786637d2194 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 713.293943] env[61907]: DEBUG nova.compute.manager [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Refreshing instance network info cache due to event network-changed-a534c8b5-7d31-46ad-accf-f786637d2194. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 713.293943] env[61907]: DEBUG oslo_concurrency.lockutils [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] Acquiring lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.386554] env[61907]: DEBUG nova.network.neutron [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.634448] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.635093] env[61907]: ERROR nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Traceback (most recent call last): [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.driver.spawn(context, instance, image_meta, [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] vm_ref = self.build_virtual_machine(instance, [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.635093] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] for vif in network_info: [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return self._sync_wrapper(fn, *args, **kwargs) [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.wait() [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self[:] = self._gt.wait() [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return self._exit_event.wait() [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] current.throw(*self._exc) [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.635429] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] result = function(*args, **kwargs) [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] return func(*args, **kwargs) [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise e [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] nwinfo = self.network_api.allocate_for_instance( [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] created_port_ids = self._update_ports_for_instance( [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] with excutils.save_and_reraise_exception(): [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] self.force_reraise() [ 713.635753] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise self.value [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] updated_port = self._update_port( [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] _ensure_no_port_binding_failure(port) [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] raise exception.PortBindingFailed(port_id=port['id']) [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] nova.exception.PortBindingFailed: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. [ 713.636091] env[61907]: ERROR nova.compute.manager [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] [ 713.636091] env[61907]: DEBUG nova.compute.utils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.636985] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.466s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.640068] env[61907]: INFO nova.compute.claims [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.643054] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Build of instance 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d was re-scheduled: Binding failed for port b3690168-eed1-45bf-90fb-d078abe58053, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 713.643492] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 713.643735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquiring lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.643883] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Acquired lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.644065] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.679609] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.681920] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 713.682199] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.683089] env[61907]: DEBUG oslo_concurrency.lockutils [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] Acquired lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.683089] env[61907]: DEBUG nova.network.neutron [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Refreshing network info cache for port a534c8b5-7d31-46ad-accf-f786637d2194 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.683741] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-730a3f40-453c-4ef7-8223-eba7a6e3de98 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.697699] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6fa725b-9a4b-4030-9672-6c7b67908d05 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.721594] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ee695d5f-1f9d-49b3-b5fa-180ea953e5ac could not be found. [ 713.721989] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.722232] env[61907]: INFO nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Took 0.04 seconds to destroy the instance on the hypervisor. [ 713.722501] env[61907]: DEBUG oslo.service.loopingcall [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.723058] env[61907]: DEBUG nova.compute.manager [-] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 713.724207] env[61907]: DEBUG nova.network.neutron [-] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.743796] env[61907]: DEBUG nova.network.neutron [-] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.889551] env[61907]: INFO nova.compute.manager [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] [instance: a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee] Took 1.02 seconds to deallocate network for instance. [ 714.163656] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.206205] env[61907]: DEBUG nova.network.neutron [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.246542] env[61907]: DEBUG nova.network.neutron [-] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.308119] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.350119] env[61907]: DEBUG nova.network.neutron [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.750486] env[61907]: INFO nova.compute.manager [-] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Took 1.03 seconds to deallocate network for instance. [ 714.753986] env[61907]: DEBUG nova.compute.claims [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 714.754195] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.813655] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Releasing lock "refresh_cache-1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.813912] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 714.814194] env[61907]: DEBUG nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 714.814396] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.831766] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.850291] env[61907]: DEBUG oslo_concurrency.lockutils [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] Releasing lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.850574] env[61907]: DEBUG nova.compute.manager [req-af8e1b78-9cda-44f2-b528-8a2a20796983 req-4a455c76-98bb-4463-a246-0f03b0542890 service nova] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Received event network-vif-deleted-a534c8b5-7d31-46ad-accf-f786637d2194 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 714.923042] env[61907]: INFO nova.scheduler.client.report [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Deleted allocations for instance a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee [ 715.015752] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718037ec-f8d3-4d2b-a1c1-32135e84d897 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.024667] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b475c291-1c7b-4f64-a5dd-6dcda08fb2c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.060172] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821221fd-2c1a-4ff4-bdbe-e7c6c9ad7a98 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.068064] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea928134-6857-46c5-a06c-4818f4c774b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.083508] env[61907]: DEBUG nova.compute.provider_tree [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.135041] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "f157c0db-054b-4d99-803c-11aa40cf0670" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.135339] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.333663] env[61907]: DEBUG nova.network.neutron [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.433299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aacd1889-61cf-4df1-95de-80fd862b3de0 tempest-InstanceActionsTestJSON-1141094721 tempest-InstanceActionsTestJSON-1141094721-project-member] Lock "a4f3e548-3111-4f4d-a3ca-b50dd2fc0fee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.564s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.587502] env[61907]: DEBUG nova.scheduler.client.report [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 715.838709] env[61907]: INFO nova.compute.manager [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] [instance: 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d] Took 1.02 seconds to deallocate network for instance. [ 715.936055] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 716.093357] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.093920] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 716.096570] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.055s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.097964] env[61907]: INFO nova.compute.claims [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.469266] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.606642] env[61907]: DEBUG nova.compute.utils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.611983] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 716.611983] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.657878] env[61907]: DEBUG nova.policy [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26bbcb77d3874106808300b1bdb36a8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c53d877a81a7422db4fdfbbc6ed37444', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.883898] env[61907]: INFO nova.scheduler.client.report [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Deleted allocations for instance 1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d [ 717.003918] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Successfully created port: 7d2bba57-849e-4d2f-84d2-2f1b052e524d {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 717.114370] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 717.395087] env[61907]: DEBUG oslo_concurrency.lockutils [None req-20f1ac15-77da-4a3f-bc6a-08de73ee2175 tempest-ServerAddressesTestJSON-1747603273 tempest-ServerAddressesTestJSON-1747603273-project-member] Lock "1c5ebe08-0b2f-4158-9eae-dd4b2dcf9a7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.003s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.483057] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad7a462-e144-4d34-9f4e-8d21f50b21b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.491112] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108b0add-c113-4ebb-9670-6ccfd9982148 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.522015] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ff88ea-a611-44f4-b489-a0a8399a3715 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.530045] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69b84d29-917d-4962-a23a-373214616787 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.545174] env[61907]: DEBUG nova.compute.provider_tree [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.900212] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 718.049425] env[61907]: DEBUG nova.scheduler.client.report [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 718.129796] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 718.156641] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 718.157154] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 718.157441] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 718.157758] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 718.161026] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 718.161026] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 718.161026] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 718.161026] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 718.161026] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 718.161247] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 718.161247] env[61907]: DEBUG nova.virt.hardware [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.161247] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a39a09-169c-4a40-8b7b-bd5670c75dbb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.168177] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bd51d6-216e-4668-abf9-3fae09446638 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.426871] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.479854] env[61907]: DEBUG nova.compute.manager [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Received event network-changed-7d2bba57-849e-4d2f-84d2-2f1b052e524d {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 718.480105] env[61907]: DEBUG nova.compute.manager [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Refreshing instance network info cache due to event network-changed-7d2bba57-849e-4d2f-84d2-2f1b052e524d. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 718.480370] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] Acquiring lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.480517] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] Acquired lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.480679] env[61907]: DEBUG nova.network.neutron [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Refreshing network info cache for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.562109] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.562109] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 718.563686] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.599s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.565056] env[61907]: INFO nova.compute.claims [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 718.773188] env[61907]: ERROR nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 718.773188] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.773188] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.773188] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.773188] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.773188] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.773188] env[61907]: ERROR nova.compute.manager raise self.value [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.773188] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.773188] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.773188] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.773656] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.773656] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.773656] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 718.773656] env[61907]: ERROR nova.compute.manager [ 718.773656] env[61907]: Traceback (most recent call last): [ 718.773656] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.773656] env[61907]: listener.cb(fileno) [ 718.773656] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.773656] env[61907]: result = function(*args, **kwargs) [ 718.773656] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.773656] env[61907]: return func(*args, **kwargs) [ 718.773656] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.773656] env[61907]: raise e [ 718.773656] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.773656] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 718.773656] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.773656] env[61907]: created_port_ids = self._update_ports_for_instance( [ 718.773656] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.773656] env[61907]: with excutils.save_and_reraise_exception(): [ 718.773656] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.773656] env[61907]: self.force_reraise() [ 718.773656] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.773656] env[61907]: raise self.value [ 718.773656] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.773656] env[61907]: updated_port = self._update_port( [ 718.773656] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.773656] env[61907]: _ensure_no_port_binding_failure(port) [ 718.773656] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.773656] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.774416] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 718.774416] env[61907]: Removing descriptor: 17 [ 718.774416] env[61907]: ERROR nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Traceback (most recent call last): [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] yield resources [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.driver.spawn(context, instance, image_meta, [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.774416] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] vm_ref = self.build_virtual_machine(instance, [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] for vif in network_info: [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self._sync_wrapper(fn, *args, **kwargs) [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.wait() [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self[:] = self._gt.wait() [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self._exit_event.wait() [ 718.774744] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] result = hub.switch() [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self.greenlet.switch() [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] result = function(*args, **kwargs) [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return func(*args, **kwargs) [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise e [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] nwinfo = self.network_api.allocate_for_instance( [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.775109] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] created_port_ids = self._update_ports_for_instance( [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] with excutils.save_and_reraise_exception(): [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.force_reraise() [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise self.value [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] updated_port = self._update_port( [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] _ensure_no_port_binding_failure(port) [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.775463] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise exception.PortBindingFailed(port_id=port['id']) [ 718.775800] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 718.775800] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] [ 718.775800] env[61907]: INFO nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Terminating instance [ 719.000879] env[61907]: DEBUG nova.network.neutron [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.064965] env[61907]: DEBUG nova.compute.utils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.066570] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 719.066857] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 719.141159] env[61907]: DEBUG nova.network.neutron [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.152176] env[61907]: DEBUG nova.policy [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd9e3b2602d342d0ba69252be126ffa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e7bd135f842743baa63fd1ac28416ad0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 719.278693] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.573605] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 719.610037] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Successfully created port: 435db216-c3ed-45b7-8397-8d3fcd34fc14 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 719.644811] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c7931e-dd3b-47de-af2d-1019dcfb3380 req-f67ca5f4-013b-4161-8d38-84ceff046393 service nova] Releasing lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.644811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.644811] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.941138] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79470dd8-3ac2-4d9d-9a29-b9b487a145fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.950156] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea16448-7846-4b59-9445-437e4a538c43 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.983387] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc608030-a207-4bae-97cd-d4743c66bee0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.991814] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a214d68d-2a6f-4d5b-971f-6f81d343318f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.007651] env[61907]: DEBUG nova.compute.provider_tree [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.173624] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.324715] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.511075] env[61907]: DEBUG nova.scheduler.client.report [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 720.523903] env[61907]: DEBUG nova.compute.manager [req-150ee6b5-d055-45fc-b996-e7328cce7bd7 req-9b75ae9e-e6c5-4c55-b0d4-09ad5de79bc9 service nova] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Received event network-vif-deleted-7d2bba57-849e-4d2f-84d2-2f1b052e524d {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 720.581703] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 720.606783] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:08:24Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='de5bdd15-ceb4-4ae8-a398-f14a0829b324',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1276338916',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 720.607005] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 720.607175] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 720.607359] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 720.607506] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 720.607647] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 720.607852] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 720.608011] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 720.608332] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 720.608502] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 720.608675] env[61907]: DEBUG nova.virt.hardware [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 720.609534] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b0e0f7f-4319-4b77-a3a5-9536cbbb9aa2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.618347] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17edb24-1328-4e14-b620-459ee74646d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.830457] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Releasing lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.831037] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 720.831240] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.831547] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e894f72-13d8-461e-8d02-b4980e1c4014 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.842491] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493e71c6-bd54-41dc-894e-d514dd410e56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.855151] env[61907]: ERROR nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 720.855151] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.855151] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.855151] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.855151] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.855151] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.855151] env[61907]: ERROR nova.compute.manager raise self.value [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.855151] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 720.855151] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.855151] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 720.855575] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.855575] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 720.855575] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 720.855575] env[61907]: ERROR nova.compute.manager [ 720.855575] env[61907]: Traceback (most recent call last): [ 720.855575] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 720.855575] env[61907]: listener.cb(fileno) [ 720.855575] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.855575] env[61907]: result = function(*args, **kwargs) [ 720.855575] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.855575] env[61907]: return func(*args, **kwargs) [ 720.855575] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.855575] env[61907]: raise e [ 720.855575] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.855575] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 720.855575] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.855575] env[61907]: created_port_ids = self._update_ports_for_instance( [ 720.855575] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.855575] env[61907]: with excutils.save_and_reraise_exception(): [ 720.855575] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.855575] env[61907]: self.force_reraise() [ 720.855575] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.855575] env[61907]: raise self.value [ 720.855575] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.855575] env[61907]: updated_port = self._update_port( [ 720.855575] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.855575] env[61907]: _ensure_no_port_binding_failure(port) [ 720.855575] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.855575] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 720.856304] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 720.856304] env[61907]: Removing descriptor: 15 [ 720.856304] env[61907]: ERROR nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] Traceback (most recent call last): [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] yield resources [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.driver.spawn(context, instance, image_meta, [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 720.856304] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] vm_ref = self.build_virtual_machine(instance, [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] vif_infos = vmwarevif.get_vif_info(self._session, [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] for vif in network_info: [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self._sync_wrapper(fn, *args, **kwargs) [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.wait() [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self[:] = self._gt.wait() [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self._exit_event.wait() [ 720.856659] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] result = hub.switch() [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self.greenlet.switch() [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] result = function(*args, **kwargs) [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return func(*args, **kwargs) [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise e [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] nwinfo = self.network_api.allocate_for_instance( [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 720.856999] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] created_port_ids = self._update_ports_for_instance( [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] with excutils.save_and_reraise_exception(): [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.force_reraise() [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise self.value [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] updated_port = self._update_port( [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] _ensure_no_port_binding_failure(port) [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 720.857374] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise exception.PortBindingFailed(port_id=port['id']) [ 720.857698] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 720.857698] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] [ 720.857698] env[61907]: INFO nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Terminating instance [ 720.868897] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 09d97063-d94f-439a-a811-b3039de1f4d7 could not be found. [ 720.869108] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.869287] env[61907]: INFO nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.869517] env[61907]: DEBUG oslo.service.loopingcall [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.869752] env[61907]: DEBUG nova.compute.manager [-] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 720.869842] env[61907]: DEBUG nova.network.neutron [-] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.885147] env[61907]: DEBUG nova.network.neutron [-] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.017667] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.018232] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 721.021736] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.142s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.024256] env[61907]: INFO nova.compute.claims [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.361046] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.361251] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquired lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.361429] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.388260] env[61907]: DEBUG nova.network.neutron [-] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.523399] env[61907]: DEBUG nova.compute.utils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.524815] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 721.884326] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.890775] env[61907]: INFO nova.compute.manager [-] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Took 1.02 seconds to deallocate network for instance. [ 721.893554] env[61907]: DEBUG nova.compute.claims [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.893729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.031391] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 722.132272] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.399054] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb4bdd3-8bb1-49c4-a8fb-f3511d103674 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.406997] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2e816c-b7c7-4b37-a7f6-3fbf80bda5e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.438648] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-356246e0-c696-4a7b-aedc-184e6547ff0b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.446402] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bff110b3-7159-43ea-98eb-ed6d7dc92498 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.461596] env[61907]: DEBUG nova.compute.provider_tree [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.592866] env[61907]: DEBUG nova.compute.manager [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Received event network-changed-435db216-c3ed-45b7-8397-8d3fcd34fc14 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 722.593056] env[61907]: DEBUG nova.compute.manager [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Refreshing instance network info cache due to event network-changed-435db216-c3ed-45b7-8397-8d3fcd34fc14. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 722.593235] env[61907]: DEBUG oslo_concurrency.lockutils [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] Acquiring lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.636302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Releasing lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.636738] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 722.636928] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 722.637243] env[61907]: DEBUG oslo_concurrency.lockutils [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] Acquired lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.637413] env[61907]: DEBUG nova.network.neutron [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Refreshing network info cache for port 435db216-c3ed-45b7-8397-8d3fcd34fc14 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 722.638717] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-354c9a56-e590-4283-8fae-457e662dfc4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.654024] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da53f95b-627c-4481-bd6a-f3357e2b7d15 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.681186] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af575382-25f6-475a-8803-48714fa0a37c could not be found. [ 722.681444] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 722.681590] env[61907]: INFO nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 722.681826] env[61907]: DEBUG oslo.service.loopingcall [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 722.682292] env[61907]: DEBUG nova.compute.manager [-] [instance: af575382-25f6-475a-8803-48714fa0a37c] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 722.682391] env[61907]: DEBUG nova.network.neutron [-] [instance: af575382-25f6-475a-8803-48714fa0a37c] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.715048] env[61907]: DEBUG nova.network.neutron [-] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.965186] env[61907]: DEBUG nova.scheduler.client.report [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 723.039201] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 723.059784] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.060053] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.060216] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.060399] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.060541] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.060686] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.060887] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.061082] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.061217] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.061374] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.061544] env[61907]: DEBUG nova.virt.hardware [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.062435] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-166f8830-edf5-4477-b5ff-cbe87146c71e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.071498] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a26ec9b9-1de2-4978-ba60-1e9c865ebd63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.085022] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 723.090525] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Creating folder: Project (4e44ef0fbac4479397739cb3b89d091c). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 723.090880] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-83322b8b-69ae-412d-9fbc-4f4992a31865 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.100660] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Created folder: Project (4e44ef0fbac4479397739cb3b89d091c) in parent group-v268168. [ 723.100846] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Creating folder: Instances. Parent ref: group-v268180. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 723.101085] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9cdb8219-6212-4ce0-aad3-cd0918d074f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.109889] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Created folder: Instances in parent group-v268180. [ 723.109986] env[61907]: DEBUG oslo.service.loopingcall [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.111032] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 723.111032] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4b1a687-4a59-4c6c-8036-b4999f99acea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.127160] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 723.127160] env[61907]: value = "task-1243609" [ 723.127160] env[61907]: _type = "Task" [ 723.127160] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.139620] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243609, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.160267] env[61907]: DEBUG nova.network.neutron [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 723.217956] env[61907]: DEBUG nova.network.neutron [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.218724] env[61907]: DEBUG nova.network.neutron [-] [instance: af575382-25f6-475a-8803-48714fa0a37c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.470908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.471642] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 723.475193] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.702s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.477290] env[61907]: INFO nova.compute.claims [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.637683] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243609, 'name': CreateVM_Task, 'duration_secs': 0.25476} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.637877] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 723.638307] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.638468] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.638806] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 723.639070] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9a6a9c5-1b6e-4165-83dd-b2c24865d9e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.643616] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 723.643616] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52af49d3-4a2f-a11e-7089-71fea7a4041c" [ 723.643616] env[61907]: _type = "Task" [ 723.643616] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.651091] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52af49d3-4a2f-a11e-7089-71fea7a4041c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.720027] env[61907]: DEBUG oslo_concurrency.lockutils [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] Releasing lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.720266] env[61907]: DEBUG nova.compute.manager [req-a4d22ded-6606-467e-92ad-3f9ad72bf52e req-f4a47293-e327-4010-b74a-05a498b3eea7 service nova] [instance: af575382-25f6-475a-8803-48714fa0a37c] Received event network-vif-deleted-435db216-c3ed-45b7-8397-8d3fcd34fc14 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 723.725413] env[61907]: INFO nova.compute.manager [-] [instance: af575382-25f6-475a-8803-48714fa0a37c] Took 1.04 seconds to deallocate network for instance. [ 723.726291] env[61907]: DEBUG nova.compute.claims [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 723.726470] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.861467] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.861754] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.983798] env[61907]: DEBUG nova.compute.utils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.986434] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 723.986535] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.041382] env[61907]: DEBUG nova.policy [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06eaf7b1ef5e4fbe97e7a00d32b105b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a73e20f371404ca0886facfc4241d6bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.154119] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52af49d3-4a2f-a11e-7089-71fea7a4041c, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.154394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.154631] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 724.154836] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 724.155253] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.155253] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 724.155423] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-260a8475-d0d8-473c-b936-0cfefe6b7bec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.163494] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 724.163639] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 724.164413] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83f0b71b-cd61-4777-90a6-f412136086e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.169763] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 724.169763] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ea251d-cc8f-03a2-80fb-1b457a745f35" [ 724.169763] env[61907]: _type = "Task" [ 724.169763] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.177534] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ea251d-cc8f-03a2-80fb-1b457a745f35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.389525] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Successfully created port: d0a29dc7-2683-4bc7-a386-8072507cc9d0 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 724.487486] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 724.681419] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ea251d-cc8f-03a2-80fb-1b457a745f35, 'name': SearchDatastore_Task, 'duration_secs': 0.009417} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.682260] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1dc72458-c31d-453b-b6ac-744c8af1c511 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.687143] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 724.687143] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a9928d-1845-24a8-fd1d-6ed8a0f6ef47" [ 724.687143] env[61907]: _type = "Task" [ 724.687143] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.694779] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a9928d-1845-24a8-fd1d-6ed8a0f6ef47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.841324] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f09797-ae8c-4b21-92ec-6a2863545b60 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.847740] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605741d6-5841-4ff6-a63c-1b9e8263ed51 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.877853] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54faee04-1062-4c8e-803f-2376af7084a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.884984] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104821ae-389c-45cf-81f6-7dd57e72aa99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.899234] env[61907]: DEBUG nova.compute.provider_tree [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.135098] env[61907]: DEBUG nova.compute.manager [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Received event network-changed-d0a29dc7-2683-4bc7-a386-8072507cc9d0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 725.135397] env[61907]: DEBUG nova.compute.manager [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Refreshing instance network info cache due to event network-changed-d0a29dc7-2683-4bc7-a386-8072507cc9d0. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 725.135397] env[61907]: DEBUG oslo_concurrency.lockutils [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] Acquiring lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.137028] env[61907]: DEBUG oslo_concurrency.lockutils [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] Acquired lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.137028] env[61907]: DEBUG nova.network.neutron [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Refreshing network info cache for port d0a29dc7-2683-4bc7-a386-8072507cc9d0 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 725.197151] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a9928d-1845-24a8-fd1d-6ed8a0f6ef47, 'name': SearchDatastore_Task, 'duration_secs': 0.009214} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.197432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.197750] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 67a7aef0-ed99-4bde-b2f0-33c36764a5d7/67a7aef0-ed99-4bde-b2f0-33c36764a5d7.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 725.198030] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131832c6-2b59-4ed7-b985-4d634c568e73 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.204528] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 725.204528] env[61907]: value = "task-1243610" [ 725.204528] env[61907]: _type = "Task" [ 725.204528] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.212693] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243610, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.316174] env[61907]: ERROR nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 725.316174] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 725.316174] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.316174] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.316174] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.316174] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.316174] env[61907]: ERROR nova.compute.manager raise self.value [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.316174] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 725.316174] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.316174] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 725.316624] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.316624] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 725.316624] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 725.316624] env[61907]: ERROR nova.compute.manager [ 725.316624] env[61907]: Traceback (most recent call last): [ 725.316624] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 725.316624] env[61907]: listener.cb(fileno) [ 725.316624] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.316624] env[61907]: result = function(*args, **kwargs) [ 725.316624] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.316624] env[61907]: return func(*args, **kwargs) [ 725.316624] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 725.316624] env[61907]: raise e [ 725.316624] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 725.316624] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 725.316624] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.316624] env[61907]: created_port_ids = self._update_ports_for_instance( [ 725.316624] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.316624] env[61907]: with excutils.save_and_reraise_exception(): [ 725.316624] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.316624] env[61907]: self.force_reraise() [ 725.316624] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.316624] env[61907]: raise self.value [ 725.316624] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.316624] env[61907]: updated_port = self._update_port( [ 725.316624] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.316624] env[61907]: _ensure_no_port_binding_failure(port) [ 725.316624] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.316624] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 725.317454] env[61907]: nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 725.317454] env[61907]: Removing descriptor: 15 [ 725.402246] env[61907]: DEBUG nova.scheduler.client.report [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 725.502562] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 725.533720] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.533994] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.534300] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.534583] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.534739] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.534886] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.535107] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.535270] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.535436] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.535600] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.535920] env[61907]: DEBUG nova.virt.hardware [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.536713] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d44b8c-aadd-4101-b351-10bb6426b841 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.547164] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa174930-b6cf-4fb6-b681-514791da9d4e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.564152] env[61907]: ERROR nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Traceback (most recent call last): [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] yield resources [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.driver.spawn(context, instance, image_meta, [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] vm_ref = self.build_virtual_machine(instance, [ 725.564152] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] for vif in network_info: [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return self._sync_wrapper(fn, *args, **kwargs) [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.wait() [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self[:] = self._gt.wait() [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return self._exit_event.wait() [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 725.564714] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] current.throw(*self._exc) [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] result = function(*args, **kwargs) [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return func(*args, **kwargs) [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise e [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] nwinfo = self.network_api.allocate_for_instance( [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] created_port_ids = self._update_ports_for_instance( [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] with excutils.save_and_reraise_exception(): [ 725.565063] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.force_reraise() [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise self.value [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] updated_port = self._update_port( [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] _ensure_no_port_binding_failure(port) [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise exception.PortBindingFailed(port_id=port['id']) [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 725.565430] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] [ 725.565430] env[61907]: INFO nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Terminating instance [ 725.654014] env[61907]: DEBUG nova.network.neutron [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.717786] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243610, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454959} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.719060] env[61907]: DEBUG nova.network.neutron [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.720672] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 67a7aef0-ed99-4bde-b2f0-33c36764a5d7/67a7aef0-ed99-4bde-b2f0-33c36764a5d7.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 725.721029] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 725.721383] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44af7043-9b14-4315-bf58-7d038890befa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.729069] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 725.729069] env[61907]: value = "task-1243611" [ 725.729069] env[61907]: _type = "Task" [ 725.729069] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.738547] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243611, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.907162] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.907728] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 725.911070] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.287s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.912196] env[61907]: INFO nova.compute.claims [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.069121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.225783] env[61907]: DEBUG oslo_concurrency.lockutils [req-5d1e8ca9-07ee-4b7e-aa7e-a44b193fe30f req-4308dbce-49d3-496a-acc9-d49f964683e8 service nova] Releasing lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.226208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquired lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.226392] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.238934] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243611, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060809} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.239194] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 726.239962] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43a3811-417b-4b83-a090-7d3961f647e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.259479] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 67a7aef0-ed99-4bde-b2f0-33c36764a5d7/67a7aef0-ed99-4bde-b2f0-33c36764a5d7.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 726.259975] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd346041-2dfa-4bc9-860c-82d29dc24ccb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.279529] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 726.279529] env[61907]: value = "task-1243612" [ 726.279529] env[61907]: _type = "Task" [ 726.279529] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.287355] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.416786] env[61907]: DEBUG nova.compute.utils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.422030] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 726.422030] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.469591] env[61907]: DEBUG nova.policy [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a925517b1c640949751bd9031c35d73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '270904dd812b40949d944a07aaee6f8d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.749278] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.789558] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243612, 'name': ReconfigVM_Task, 'duration_secs': 0.264594} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.789872] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 67a7aef0-ed99-4bde-b2f0-33c36764a5d7/67a7aef0-ed99-4bde-b2f0-33c36764a5d7.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 726.790554] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19303231-defd-496a-8c65-4618a2772052 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.797383] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 726.797383] env[61907]: value = "task-1243613" [ 726.797383] env[61907]: _type = "Task" [ 726.797383] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.805349] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243613, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.833653] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.851951] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Successfully created port: a6572b46-5a9b-464a-b414-90a03db283d7 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 726.920125] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 727.160087] env[61907]: DEBUG nova.compute.manager [req-32eaa83d-57e9-4ac2-bc6d-0f21a963230d req-19e395aa-ff54-4974-b01a-7de9b4049354 service nova] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Received event network-vif-deleted-d0a29dc7-2683-4bc7-a386-8072507cc9d0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 727.280812] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1440a44-65ee-43c9-bd88-e5d4110849a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.289294] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0e765d-bdf1-4d1f-bb6d-b680d89731be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.324386] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f7ae68-72d9-40eb-b237-caa9abd8d0d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.334843] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243613, 'name': Rename_Task, 'duration_secs': 0.131725} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.336833] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 727.337424] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Releasing lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.337760] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 727.337950] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.338585] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4227c220-e11f-4ce1-bcde-262899967a96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.340234] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d143ca78-f146-4ad4-a209-dcaa1a197fc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.346493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9b17b0-8b77-46e9-a343-884de4a3201d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.352637] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 727.352637] env[61907]: value = "task-1243614" [ 727.352637] env[61907]: _type = "Task" [ 727.352637] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.364057] env[61907]: DEBUG nova.compute.provider_tree [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.369789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b69aac8-835a-454d-9c4d-7e0382a530e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.382973] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243614, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.394407] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3 could not be found. [ 727.394623] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.394802] env[61907]: INFO nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Took 0.06 seconds to destroy the instance on the hypervisor. [ 727.395048] env[61907]: DEBUG oslo.service.loopingcall [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.395256] env[61907]: DEBUG nova.compute.manager [-] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 727.395352] env[61907]: DEBUG nova.network.neutron [-] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.422061] env[61907]: DEBUG nova.network.neutron [-] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.863006] env[61907]: DEBUG oslo_vmware.api [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243614, 'name': PowerOnVM_Task, 'duration_secs': 0.414708} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.863307] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 727.863506] env[61907]: INFO nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Took 4.82 seconds to spawn the instance on the hypervisor. [ 727.863680] env[61907]: DEBUG nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 727.864664] env[61907]: ERROR nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 727.864664] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.864664] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.864664] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.864664] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.864664] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.864664] env[61907]: ERROR nova.compute.manager raise self.value [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.864664] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 727.864664] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.864664] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 727.865252] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.865252] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 727.865252] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 727.865252] env[61907]: ERROR nova.compute.manager [ 727.865252] env[61907]: Traceback (most recent call last): [ 727.865252] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 727.865252] env[61907]: listener.cb(fileno) [ 727.865252] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.865252] env[61907]: result = function(*args, **kwargs) [ 727.865252] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.865252] env[61907]: return func(*args, **kwargs) [ 727.865252] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 727.865252] env[61907]: raise e [ 727.865252] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.865252] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 727.865252] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.865252] env[61907]: created_port_ids = self._update_ports_for_instance( [ 727.865252] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.865252] env[61907]: with excutils.save_and_reraise_exception(): [ 727.865252] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.865252] env[61907]: self.force_reraise() [ 727.865252] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.865252] env[61907]: raise self.value [ 727.865252] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.865252] env[61907]: updated_port = self._update_port( [ 727.865252] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.865252] env[61907]: _ensure_no_port_binding_failure(port) [ 727.865252] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.865252] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 727.866124] env[61907]: nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 727.866124] env[61907]: Removing descriptor: 15 [ 727.866124] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8040fc-85b5-4459-845e-5f2721f4cf67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.868547] env[61907]: DEBUG nova.scheduler.client.report [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 727.924173] env[61907]: DEBUG nova.network.neutron [-] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.930427] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 727.961247] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 727.961511] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 727.961663] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 727.961842] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 727.961986] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 727.962151] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 727.962352] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 727.962505] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 727.962664] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 727.962818] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 727.962986] env[61907]: DEBUG nova.virt.hardware [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 727.964026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b5f13b-fe56-4706-bac9-508b078882fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.971767] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ae98ff-2dec-4131-8366-e7ab3a567800 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.986020] env[61907]: ERROR nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Traceback (most recent call last): [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] yield resources [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.driver.spawn(context, instance, image_meta, [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self._vmops.spawn(context, instance, image_meta, injected_files, [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] vm_ref = self.build_virtual_machine(instance, [ 727.986020] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] vif_infos = vmwarevif.get_vif_info(self._session, [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] for vif in network_info: [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return self._sync_wrapper(fn, *args, **kwargs) [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.wait() [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self[:] = self._gt.wait() [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return self._exit_event.wait() [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 727.986427] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] current.throw(*self._exc) [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] result = function(*args, **kwargs) [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return func(*args, **kwargs) [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise e [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] nwinfo = self.network_api.allocate_for_instance( [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] created_port_ids = self._update_ports_for_instance( [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] with excutils.save_and_reraise_exception(): [ 727.986783] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.force_reraise() [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise self.value [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] updated_port = self._update_port( [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] _ensure_no_port_binding_failure(port) [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise exception.PortBindingFailed(port_id=port['id']) [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 727.987250] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] [ 727.987250] env[61907]: INFO nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Terminating instance [ 728.372945] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.373538] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 728.376504] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.823s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.386454] env[61907]: INFO nova.compute.manager [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Took 24.43 seconds to build instance. [ 728.427590] env[61907]: INFO nova.compute.manager [-] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Took 1.03 seconds to deallocate network for instance. [ 728.429937] env[61907]: DEBUG nova.compute.claims [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.430130] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.493455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquiring lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.493652] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquired lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.493835] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.881199] env[61907]: DEBUG nova.compute.utils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.882674] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 728.882850] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.894808] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8df9bdcf-c88f-4198-8057-2f9f4e4ddec5 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.203s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.981566] env[61907]: DEBUG nova.policy [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '06eaf7b1ef5e4fbe97e7a00d32b105b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a73e20f371404ca0886facfc4241d6bd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.037828] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.173116] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.325272] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfbfba7-70f0-438f-bb01-03f6579ceb50 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.334292] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b8177e-3882-4874-98bf-207c25ec4190 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.367848] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24c01830-ae69-4aa8-bc73-2829deeabb69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.374087] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aaa370c-bd69-46d6-acbf-fded9fd5fdde {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.388135] env[61907]: DEBUG nova.compute.provider_tree [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.395242] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 729.397409] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 729.401610] env[61907]: DEBUG nova.compute.manager [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Received event network-changed-a6572b46-5a9b-464a-b414-90a03db283d7 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 729.401801] env[61907]: DEBUG nova.compute.manager [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Refreshing instance network info cache due to event network-changed-a6572b46-5a9b-464a-b414-90a03db283d7. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 729.401986] env[61907]: DEBUG oslo_concurrency.lockutils [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] Acquiring lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.484185] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Successfully created port: fe20272c-cdd7-495c-818b-12ecb67f7ac2 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.676055] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Releasing lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.676652] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 729.676932] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.677387] env[61907]: DEBUG oslo_concurrency.lockutils [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] Acquired lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.677644] env[61907]: DEBUG nova.network.neutron [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Refreshing network info cache for port a6572b46-5a9b-464a-b414-90a03db283d7 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.678997] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4396b4aa-3c4d-4266-a10e-9df3e6b1ab2e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.691840] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b083fa6-c81a-4f74-881c-da9fc6c9373a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.726159] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827 could not be found. [ 729.726159] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.726159] env[61907]: INFO nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Took 0.05 seconds to destroy the instance on the hypervisor. [ 729.726386] env[61907]: DEBUG oslo.service.loopingcall [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.726691] env[61907]: DEBUG nova.compute.manager [-] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 729.726821] env[61907]: DEBUG nova.network.neutron [-] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.752021] env[61907]: DEBUG nova.network.neutron [-] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.892685] env[61907]: DEBUG nova.scheduler.client.report [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 729.933353] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.024121] env[61907]: DEBUG nova.compute.manager [None req-bbb6a488-ea65-4c89-971f-50b4bccf5a82 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 730.024121] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81bcd9b-1720-4c1f-a01a-988af90ed0ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.136497] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.136783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.136998] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.137306] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.137384] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.139589] env[61907]: INFO nova.compute.manager [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Terminating instance [ 730.213501] env[61907]: DEBUG nova.network.neutron [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.253454] env[61907]: DEBUG nova.network.neutron [-] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.350276] env[61907]: DEBUG nova.network.neutron [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.399279] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.401773] env[61907]: ERROR nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Traceback (most recent call last): [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.driver.spawn(context, instance, image_meta, [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] vm_ref = self.build_virtual_machine(instance, [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.401773] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] for vif in network_info: [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self._sync_wrapper(fn, *args, **kwargs) [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.wait() [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self[:] = self._gt.wait() [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self._exit_event.wait() [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] result = hub.switch() [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.402223] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return self.greenlet.switch() [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] result = function(*args, **kwargs) [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] return func(*args, **kwargs) [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise e [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] nwinfo = self.network_api.allocate_for_instance( [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] created_port_ids = self._update_ports_for_instance( [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] with excutils.save_and_reraise_exception(): [ 730.402659] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] self.force_reraise() [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise self.value [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] updated_port = self._update_port( [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] _ensure_no_port_binding_failure(port) [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] raise exception.PortBindingFailed(port_id=port['id']) [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] nova.exception.PortBindingFailed: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. [ 730.403103] env[61907]: ERROR nova.compute.manager [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] [ 730.403494] env[61907]: DEBUG nova.compute.utils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.403494] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.648s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.405010] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Build of instance 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5 was re-scheduled: Binding failed for port c801028f-2d97-4011-bdf5-cd27de91e138, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 730.405470] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 730.405690] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquiring lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.405839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Acquired lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.406013] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.412186] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 730.446106] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.446382] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.446532] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.446719] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.446869] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.450222] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.450978] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.450978] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.451236] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.451424] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.451704] env[61907]: DEBUG nova.virt.hardware [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.452825] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa471687-c34f-438c-a72c-040893665641 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.461363] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605c7b26-15e4-4d9b-9788-d3b311325e69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.537888] env[61907]: INFO nova.compute.manager [None req-bbb6a488-ea65-4c89-971f-50b4bccf5a82 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] instance snapshotting [ 730.538541] env[61907]: DEBUG nova.objects.instance [None req-bbb6a488-ea65-4c89-971f-50b4bccf5a82 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lazy-loading 'flavor' on Instance uuid 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.643471] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.643693] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquired lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.643872] env[61907]: DEBUG nova.network.neutron [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.757621] env[61907]: INFO nova.compute.manager [-] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Took 1.03 seconds to deallocate network for instance. [ 730.761682] env[61907]: DEBUG nova.compute.claims [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 730.761832] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.853139] env[61907]: DEBUG oslo_concurrency.lockutils [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] Releasing lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.853449] env[61907]: DEBUG nova.compute.manager [req-c02acd7e-5271-43ec-aaa5-134532d84be6 req-96f44dfa-51a5-48bb-9ed3-2f1774462f80 service nova] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Received event network-vif-deleted-a6572b46-5a9b-464a-b414-90a03db283d7 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 730.940352] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.970121] env[61907]: ERROR nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 730.970121] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.970121] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.970121] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.970121] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.970121] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.970121] env[61907]: ERROR nova.compute.manager raise self.value [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.970121] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 730.970121] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.970121] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 730.970577] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.970577] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 730.970577] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 730.970577] env[61907]: ERROR nova.compute.manager [ 730.970577] env[61907]: Traceback (most recent call last): [ 730.970577] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 730.970577] env[61907]: listener.cb(fileno) [ 730.970577] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.970577] env[61907]: result = function(*args, **kwargs) [ 730.970577] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.970577] env[61907]: return func(*args, **kwargs) [ 730.970577] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.970577] env[61907]: raise e [ 730.970577] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.970577] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 730.970577] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.970577] env[61907]: created_port_ids = self._update_ports_for_instance( [ 730.970577] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.970577] env[61907]: with excutils.save_and_reraise_exception(): [ 730.970577] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.970577] env[61907]: self.force_reraise() [ 730.970577] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.970577] env[61907]: raise self.value [ 730.970577] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.970577] env[61907]: updated_port = self._update_port( [ 730.970577] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.970577] env[61907]: _ensure_no_port_binding_failure(port) [ 730.970577] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.970577] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 730.971408] env[61907]: nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 730.971408] env[61907]: Removing descriptor: 15 [ 730.971408] env[61907]: ERROR nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Traceback (most recent call last): [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] yield resources [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.driver.spawn(context, instance, image_meta, [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.971408] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] vm_ref = self.build_virtual_machine(instance, [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] for vif in network_info: [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self._sync_wrapper(fn, *args, **kwargs) [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.wait() [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self[:] = self._gt.wait() [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self._exit_event.wait() [ 730.971735] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] result = hub.switch() [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self.greenlet.switch() [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] result = function(*args, **kwargs) [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return func(*args, **kwargs) [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise e [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] nwinfo = self.network_api.allocate_for_instance( [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.972089] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] created_port_ids = self._update_ports_for_instance( [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] with excutils.save_and_reraise_exception(): [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.force_reraise() [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise self.value [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] updated_port = self._update_port( [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] _ensure_no_port_binding_failure(port) [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.972430] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise exception.PortBindingFailed(port_id=port['id']) [ 730.972737] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 730.972737] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] [ 730.972737] env[61907]: INFO nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Terminating instance [ 731.046336] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32553d0-a506-4671-a626-38d75bfcfd42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.074724] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.079342] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60b09d23-476d-4482-b7eb-5c11d4e021d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.170630] env[61907]: DEBUG nova.network.neutron [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.230533] env[61907]: DEBUG nova.network.neutron [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.303104] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ca8a95-a76d-4ceb-94d7-e16d3358c1e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.310846] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad9f6b0-7663-419c-a838-f3efa15aa9d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.340206] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e45b5b70-8e15-4839-a2d8-682b4ace4fea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.347534] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54cbcd6-3fa3-4efd-b01d-ff8ef0cafd71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.360364] env[61907]: DEBUG nova.compute.provider_tree [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.422235] env[61907]: DEBUG nova.compute.manager [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Received event network-changed-fe20272c-cdd7-495c-818b-12ecb67f7ac2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 731.422587] env[61907]: DEBUG nova.compute.manager [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Refreshing instance network info cache due to event network-changed-fe20272c-cdd7-495c-818b-12ecb67f7ac2. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 731.422702] env[61907]: DEBUG oslo_concurrency.lockutils [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] Acquiring lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.422784] env[61907]: DEBUG oslo_concurrency.lockutils [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] Acquired lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.422940] env[61907]: DEBUG nova.network.neutron [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Refreshing network info cache for port fe20272c-cdd7-495c-818b-12ecb67f7ac2 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.476807] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.582945] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Releasing lock "refresh_cache-81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.583225] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 731.583472] env[61907]: DEBUG nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 731.583675] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.590060] env[61907]: DEBUG nova.compute.manager [None req-bbb6a488-ea65-4c89-971f-50b4bccf5a82 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance disappeared during snapshot {{(pid=61907) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 731.600517] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.734505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Releasing lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.735052] env[61907]: DEBUG nova.compute.manager [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 731.735140] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 731.736821] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2343824-6105-4ba2-97b1-3eb0ae6f0ca9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.745258] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 731.745258] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f18d284c-0a9e-4d6d-95b2-1e7267d3ba9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.751817] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 731.751817] env[61907]: value = "task-1243615" [ 731.751817] env[61907]: _type = "Task" [ 731.751817] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.760166] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.774237] env[61907]: DEBUG nova.compute.manager [None req-bbb6a488-ea65-4c89-971f-50b4bccf5a82 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Found 0 images (rotation: 2) {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 731.864030] env[61907]: DEBUG nova.scheduler.client.report [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 731.942885] env[61907]: DEBUG nova.network.neutron [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.043525] env[61907]: DEBUG nova.network.neutron [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.102709] env[61907]: DEBUG nova.network.neutron [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.262066] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243615, 'name': PowerOffVM_Task, 'duration_secs': 0.268396} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.262066] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 732.262066] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 732.262066] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4f5f386-bc5e-4c81-9cc0-3f0a50daf5a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.283638] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 732.283862] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 732.284057] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Deleting the datastore file [datastore2] 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 732.284314] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6b6af87b-5684-4627-964b-5f15cf0b65af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.289896] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for the task: (returnval){ [ 732.289896] env[61907]: value = "task-1243617" [ 732.289896] env[61907]: _type = "Task" [ 732.289896] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.297341] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243617, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.368766] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.967s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.369439] env[61907]: ERROR nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Traceback (most recent call last): [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.driver.spawn(context, instance, image_meta, [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] vm_ref = self.build_virtual_machine(instance, [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.369439] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] for vif in network_info: [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self._sync_wrapper(fn, *args, **kwargs) [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.wait() [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self[:] = self._gt.wait() [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self._exit_event.wait() [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] result = hub.switch() [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 732.369802] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return self.greenlet.switch() [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] result = function(*args, **kwargs) [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] return func(*args, **kwargs) [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise e [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] nwinfo = self.network_api.allocate_for_instance( [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] created_port_ids = self._update_ports_for_instance( [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] with excutils.save_and_reraise_exception(): [ 732.370192] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] self.force_reraise() [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise self.value [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] updated_port = self._update_port( [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] _ensure_no_port_binding_failure(port) [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] raise exception.PortBindingFailed(port_id=port['id']) [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] nova.exception.PortBindingFailed: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. [ 732.370593] env[61907]: ERROR nova.compute.manager [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] [ 732.370905] env[61907]: DEBUG nova.compute.utils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.371410] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.902s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.372835] env[61907]: INFO nova.compute.claims [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.376394] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Build of instance ee695d5f-1f9d-49b3-b5fa-180ea953e5ac was re-scheduled: Binding failed for port a534c8b5-7d31-46ad-accf-f786637d2194, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 732.376841] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 732.377107] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.377261] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.377419] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.546835] env[61907]: DEBUG oslo_concurrency.lockutils [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] Releasing lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.547205] env[61907]: DEBUG nova.compute.manager [req-cc0c0bde-1e5b-4c13-9790-48340d161703 req-b4b81c51-dcd8-4e93-af5c-566f0896bb48 service nova] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Received event network-vif-deleted-fe20272c-cdd7-495c-818b-12ecb67f7ac2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 732.547513] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquired lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.547683] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.605804] env[61907]: INFO nova.compute.manager [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] [instance: 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5] Took 1.02 seconds to deallocate network for instance. [ 732.799524] env[61907]: DEBUG oslo_vmware.api [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Task: {'id': task-1243617, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080511} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.799756] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 732.800009] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 732.800209] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 732.800384] env[61907]: INFO nova.compute.manager [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Took 1.07 seconds to destroy the instance on the hypervisor. [ 732.800620] env[61907]: DEBUG oslo.service.loopingcall [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 732.800804] env[61907]: DEBUG nova.compute.manager [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 732.800901] env[61907]: DEBUG nova.network.neutron [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 732.816686] env[61907]: DEBUG nova.network.neutron [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.904020] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.990151] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.072646] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.122337] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.319770] env[61907]: DEBUG nova.network.neutron [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.494156] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "refresh_cache-ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.494411] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 733.494588] env[61907]: DEBUG nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 733.494751] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.509232] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.625769] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Releasing lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.626198] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 733.626425] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.626816] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-124bbfd5-4943-4bff-952b-38059b2d7e25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.631500] env[61907]: INFO nova.scheduler.client.report [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Deleted allocations for instance 81ab83b9-26ba-4dbd-a39a-11c05a8c90d5 [ 733.645130] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ba83d3-b9e2-44b5-93e5-5dbfa51fd71e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.669049] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ea6d2b19-be25-4da3-b8af-11c3dca04df8 could not be found. [ 733.669300] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 733.669478] env[61907]: INFO nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 733.669715] env[61907]: DEBUG oslo.service.loopingcall [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.672184] env[61907]: DEBUG nova.compute.manager [-] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 733.672290] env[61907]: DEBUG nova.network.neutron [-] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.685497] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c66ca76e-0e6e-4b3f-8ae4-ffca9f8ecd1a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.689298] env[61907]: DEBUG nova.network.neutron [-] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.694169] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8d2461-f9ab-474e-8e98-4074d8082555 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.725796] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45740212-bbf3-4c53-93a6-46e425f1b616 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.733555] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccfe0de-8aa9-45f9-89fb-eb4733c58a12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.747431] env[61907]: DEBUG nova.compute.provider_tree [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.821350] env[61907]: INFO nova.compute.manager [-] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Took 1.02 seconds to deallocate network for instance. [ 734.011963] env[61907]: DEBUG nova.network.neutron [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.140295] env[61907]: DEBUG oslo_concurrency.lockutils [None req-22bd7647-0a3b-4723-8460-9cfd0e42aa4f tempest-ServerTagsTestJSON-1615128928 tempest-ServerTagsTestJSON-1615128928-project-member] Lock "81ab83b9-26ba-4dbd-a39a-11c05a8c90d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.832s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.194599] env[61907]: DEBUG nova.network.neutron [-] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.250202] env[61907]: DEBUG nova.scheduler.client.report [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 734.327809] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.514856] env[61907]: INFO nova.compute.manager [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: ee695d5f-1f9d-49b3-b5fa-180ea953e5ac] Took 1.02 seconds to deallocate network for instance. [ 734.642927] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 734.697370] env[61907]: INFO nova.compute.manager [-] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Took 1.02 seconds to deallocate network for instance. [ 734.699425] env[61907]: DEBUG nova.compute.claims [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 734.699602] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.755253] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.755766] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 734.758222] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.331s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.759693] env[61907]: INFO nova.compute.claims [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.117430] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.117671] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.168351] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.265205] env[61907]: DEBUG nova.compute.utils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.268165] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 735.553835] env[61907]: INFO nova.scheduler.client.report [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Deleted allocations for instance ee695d5f-1f9d-49b3-b5fa-180ea953e5ac [ 735.625835] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 735.626056] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 735.626603] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 735.769220] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 736.062769] env[61907]: DEBUG oslo_concurrency.lockutils [None req-60df3971-2feb-4606-a029-9894914f3136 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "ee695d5f-1f9d-49b3-b5fa-180ea953e5ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.291s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.065639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddf1dfb2-8344-47a1-ac62-e6aa83dc2273 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.074766] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953e2865-fcee-491c-ab32-2c7ce2bb771a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.105053] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ddfeb5-58dd-4252-9e6d-9d95f0daab60 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.114279] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d675522-afad-46ea-b893-8700ea45ebd6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.127999] env[61907]: DEBUG nova.compute.provider_tree [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.132401] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.132586] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: af575382-25f6-475a-8803-48714fa0a37c] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.132788] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.132959] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.133156] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.133329] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.133493] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 736.148193] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.148331] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.148469] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 736.148620] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 736.569171] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 736.630494] env[61907]: DEBUG nova.scheduler.client.report [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 736.783286] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 736.810099] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.810387] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.810545] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.811153] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.811153] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.811153] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.811295] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.811423] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.811541] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.811729] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.811865] env[61907]: DEBUG nova.virt.hardware [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.812739] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3238006-e6e8-46f0-9e68-90c9e09b7573 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.820570] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b3ddd96-e6ed-47a4-90eb-268552533086 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.833766] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.839240] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Creating folder: Project (064ee37ef4f74392b5ae3dea5498876a). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.839516] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1fcec2fa-8809-4b93-b83f-81aaf868eeae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.850538] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Created folder: Project (064ee37ef4f74392b5ae3dea5498876a) in parent group-v268168. [ 736.851330] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Creating folder: Instances. Parent ref: group-v268183. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.851330] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03e05a2f-1947-417d-bcd0-77538745003e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.859921] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Created folder: Instances in parent group-v268183. [ 736.860170] env[61907]: DEBUG oslo.service.loopingcall [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.860340] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 736.860526] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-decf0050-1e8d-40df-a5d3-b264043e591e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.877168] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.877168] env[61907]: value = "task-1243620" [ 736.877168] env[61907]: _type = "Task" [ 736.877168] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.884541] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243620, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.902291] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "e5d4890e-be1b-4b28-9721-89ef678c1182" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.903022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.089026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.136192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.136919] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 737.139670] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.246s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.172373] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.386398] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243620, 'name': CreateVM_Task, 'duration_secs': 0.260325} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.386571] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.386985] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.387163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.387477] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.387726] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4df482b-5b18-4486-b59e-2b639fe3f574 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.392277] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 737.392277] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b11e86-b296-16fc-7ed4-3b5a15954501" [ 737.392277] env[61907]: _type = "Task" [ 737.392277] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.399644] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b11e86-b296-16fc-7ed4-3b5a15954501, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.653465] env[61907]: DEBUG nova.compute.utils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 737.655041] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 737.655214] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 737.740303] env[61907]: DEBUG nova.policy [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9edf784d907947d4b873973c3188d2a9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'abfbfd1cc886447fac7c0af3ab1c5867', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 737.772973] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.902517] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b11e86-b296-16fc-7ed4-3b5a15954501, 'name': SearchDatastore_Task, 'duration_secs': 0.011525} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.905016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.905295] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.905525] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.905669] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.905846] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 737.906474] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96fbadfc-0409-47b6-b311-f2d75b0e72e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.914550] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 737.914660] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 737.915403] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49fc11ce-03f8-4fa7-96b9-458b7f3b093f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.923630] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 737.923630] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5280d647-b56c-5daf-bc26-8a0f604a9f86" [ 737.923630] env[61907]: _type = "Task" [ 737.923630] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.931878] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5280d647-b56c-5daf-bc26-8a0f604a9f86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.016052] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff7e5cd-d809-4996-b7f2-4dad541c0fbe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.023807] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be52ae9-fa48-4a15-985c-a7c65afc0d88 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.052910] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e6ccd3-6ccc-41f1-af97-b6a63d1e550c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.059779] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77d09c1-b0ed-497b-a3fd-fbcc06680f40 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.072722] env[61907]: DEBUG nova.compute.provider_tree [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.096575] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Successfully created port: 38a00017-9f7b-4899-8e16-3e432ebb3cf1 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 738.158927] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 738.278314] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-67a7aef0-ed99-4bde-b2f0-33c36764a5d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.278314] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 738.278314] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.278314] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.278314] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.278314] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.278557] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.280027] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.280114] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 738.280248] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.441689] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5280d647-b56c-5daf-bc26-8a0f604a9f86, 'name': SearchDatastore_Task, 'duration_secs': 0.008267} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.441689] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07e6961b-1768-4279-aa56-ec524f32ca97 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.446754] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 738.446754] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f33800-0cd7-0f1d-6895-900264336202" [ 738.446754] env[61907]: _type = "Task" [ 738.446754] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.455771] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f33800-0cd7-0f1d-6895-900264336202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.576352] env[61907]: DEBUG nova.scheduler.client.report [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 738.785805] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.902396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "5c908f9c-7efb-4588-b6bf-70babb173767" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.902698] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.956823] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f33800-0cd7-0f1d-6895-900264336202, 'name': SearchDatastore_Task, 'duration_secs': 0.013217} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.957116] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.957374] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 738.957632] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5d29594d-37a3-4793-bb70-6c25e07bf7ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.965202] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 738.965202] env[61907]: value = "task-1243621" [ 738.965202] env[61907]: _type = "Task" [ 738.965202] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.972564] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.031631] env[61907]: DEBUG nova.compute.manager [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Received event network-changed-38a00017-9f7b-4899-8e16-3e432ebb3cf1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 739.031854] env[61907]: DEBUG nova.compute.manager [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Refreshing instance network info cache due to event network-changed-38a00017-9f7b-4899-8e16-3e432ebb3cf1. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 739.032080] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] Acquiring lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.032227] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] Acquired lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.032415] env[61907]: DEBUG nova.network.neutron [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Refreshing network info cache for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 739.086316] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.947s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.086947] env[61907]: ERROR nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Traceback (most recent call last): [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.driver.spawn(context, instance, image_meta, [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] vm_ref = self.build_virtual_machine(instance, [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.086947] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] for vif in network_info: [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self._sync_wrapper(fn, *args, **kwargs) [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.wait() [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self[:] = self._gt.wait() [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self._exit_event.wait() [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] result = hub.switch() [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 739.087310] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return self.greenlet.switch() [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] result = function(*args, **kwargs) [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] return func(*args, **kwargs) [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise e [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] nwinfo = self.network_api.allocate_for_instance( [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] created_port_ids = self._update_ports_for_instance( [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] with excutils.save_and_reraise_exception(): [ 739.087704] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] self.force_reraise() [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise self.value [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] updated_port = self._update_port( [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] _ensure_no_port_binding_failure(port) [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] raise exception.PortBindingFailed(port_id=port['id']) [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] nova.exception.PortBindingFailed: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. [ 739.088286] env[61907]: ERROR nova.compute.manager [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] [ 739.088673] env[61907]: DEBUG nova.compute.utils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 739.088897] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.362s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.092033] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Build of instance 09d97063-d94f-439a-a811-b3039de1f4d7 was re-scheduled: Binding failed for port 7d2bba57-849e-4d2f-84d2-2f1b052e524d, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 739.092281] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 739.093027] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.093027] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.093027] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 739.172020] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 739.190848] env[61907]: ERROR nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 739.190848] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.190848] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.190848] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.190848] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.190848] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.190848] env[61907]: ERROR nova.compute.manager raise self.value [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.190848] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 739.190848] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.190848] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 739.191691] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.191691] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 739.191691] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 739.191691] env[61907]: ERROR nova.compute.manager [ 739.191691] env[61907]: Traceback (most recent call last): [ 739.191691] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 739.191691] env[61907]: listener.cb(fileno) [ 739.191691] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.191691] env[61907]: result = function(*args, **kwargs) [ 739.191691] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.191691] env[61907]: return func(*args, **kwargs) [ 739.191691] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.191691] env[61907]: raise e [ 739.191691] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.191691] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 739.191691] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.191691] env[61907]: created_port_ids = self._update_ports_for_instance( [ 739.191691] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.191691] env[61907]: with excutils.save_and_reraise_exception(): [ 739.191691] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.191691] env[61907]: self.force_reraise() [ 739.191691] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.191691] env[61907]: raise self.value [ 739.191691] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.191691] env[61907]: updated_port = self._update_port( [ 739.191691] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.191691] env[61907]: _ensure_no_port_binding_failure(port) [ 739.191691] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.191691] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 739.192722] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 739.192722] env[61907]: Removing descriptor: 17 [ 739.200411] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 739.200665] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 739.200821] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 739.201050] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 739.201227] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 739.201377] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 739.201588] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 739.201745] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 739.201909] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 739.202090] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 739.202280] env[61907]: DEBUG nova.virt.hardware [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 739.203802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ea2aa5-5731-48f2-8119-4601eac7ca2b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.212044] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04018f37-4d0b-49d2-bf19-0f62a08aef1f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.230162] env[61907]: ERROR nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Traceback (most recent call last): [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] yield resources [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.driver.spawn(context, instance, image_meta, [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] vm_ref = self.build_virtual_machine(instance, [ 739.230162] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] for vif in network_info: [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return self._sync_wrapper(fn, *args, **kwargs) [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.wait() [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self[:] = self._gt.wait() [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return self._exit_event.wait() [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 739.230747] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] current.throw(*self._exc) [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] result = function(*args, **kwargs) [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return func(*args, **kwargs) [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise e [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] nwinfo = self.network_api.allocate_for_instance( [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] created_port_ids = self._update_ports_for_instance( [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] with excutils.save_and_reraise_exception(): [ 739.231193] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.force_reraise() [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise self.value [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] updated_port = self._update_port( [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] _ensure_no_port_binding_failure(port) [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise exception.PortBindingFailed(port_id=port['id']) [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 739.231617] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] [ 739.231617] env[61907]: INFO nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Terminating instance [ 739.475806] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.443876} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.476087] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 739.476738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.476738] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7b2839bd-b6e3-4fb7-85e2-d03c2bab953f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.483099] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 739.483099] env[61907]: value = "task-1243622" [ 739.483099] env[61907]: _type = "Task" [ 739.483099] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.492128] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243622, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.552029] env[61907]: DEBUG nova.network.neutron [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.735271] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.737122] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquiring lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.873680] env[61907]: DEBUG nova.network.neutron [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.876189] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.922442] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bd62dd-b73c-477e-9a7d-67b4da2981a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.930686] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85647a94-db7b-4d51-a9bd-8ed5ebbce772 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.960969] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd970c9-80a5-41ac-8776-cb4d9b777c9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.968366] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e3edb3-ea04-417b-8354-7943f149924f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.981250] env[61907]: DEBUG nova.compute.provider_tree [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.990581] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243622, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.311551} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.991398] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 739.992283] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68588dbb-269c-4184-86a3-25332acbccde {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.012673] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.013092] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8995df88-07c2-48c6-b05b-79f9b9632d67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.032824] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 740.032824] env[61907]: value = "task-1243623" [ 740.032824] env[61907]: _type = "Task" [ 740.032824] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.040137] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.378413] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaea1d6d-bf2b-4c19-b7d9-d41f1d9c82d5 req-d1e0c01a-2edb-435f-a68c-93ab8eab1e0e service nova] Releasing lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.378954] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquired lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.379239] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 740.386108] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Releasing lock "refresh_cache-09d97063-d94f-439a-a811-b3039de1f4d7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.386322] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 740.386651] env[61907]: DEBUG nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 740.386871] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 740.401883] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.487799] env[61907]: DEBUG nova.scheduler.client.report [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 740.542278] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243623, 'name': ReconfigVM_Task, 'duration_secs': 0.255107} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.542550] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfigured VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.543268] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0733bbad-fc99-4b94-8e14-3c54e0f25340 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.548762] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 740.548762] env[61907]: value = "task-1243624" [ 740.548762] env[61907]: _type = "Task" [ 740.548762] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.556092] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243624, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.897813] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.904125] env[61907]: DEBUG nova.network.neutron [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.993917] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.904s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.993917] env[61907]: ERROR nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] Traceback (most recent call last): [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.driver.spawn(context, instance, image_meta, [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 740.993917] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] vm_ref = self.build_virtual_machine(instance, [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] vif_infos = vmwarevif.get_vif_info(self._session, [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] for vif in network_info: [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self._sync_wrapper(fn, *args, **kwargs) [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.wait() [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self[:] = self._gt.wait() [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self._exit_event.wait() [ 740.994394] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] result = hub.switch() [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return self.greenlet.switch() [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] result = function(*args, **kwargs) [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] return func(*args, **kwargs) [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise e [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] nwinfo = self.network_api.allocate_for_instance( [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 740.994729] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] created_port_ids = self._update_ports_for_instance( [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] with excutils.save_and_reraise_exception(): [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] self.force_reraise() [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise self.value [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] updated_port = self._update_port( [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] _ensure_no_port_binding_failure(port) [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 740.995083] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] raise exception.PortBindingFailed(port_id=port['id']) [ 740.995398] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] nova.exception.PortBindingFailed: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. [ 740.995398] env[61907]: ERROR nova.compute.manager [instance: af575382-25f6-475a-8803-48714fa0a37c] [ 740.995398] env[61907]: DEBUG nova.compute.utils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 740.996680] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.997882] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Build of instance af575382-25f6-475a-8803-48714fa0a37c was re-scheduled: Binding failed for port 435db216-c3ed-45b7-8397-8d3fcd34fc14, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 740.998368] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 740.998633] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquiring lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.998880] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Acquired lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.999101] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 741.000045] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.570s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.060026] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243624, 'name': Rename_Task, 'duration_secs': 0.13017} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.061147] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 741.062200] env[61907]: DEBUG nova.compute.manager [req-4e41d9dd-851b-4ae2-bea6-cf0d47577e91 req-2da6ea3b-fea7-4538-ab49-29eedc0425bc service nova] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Received event network-vif-deleted-38a00017-9f7b-4899-8e16-3e432ebb3cf1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 741.062491] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c29d4617-a8c9-4018-9b9f-203040732426 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.068479] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 741.068479] env[61907]: value = "task-1243625" [ 741.068479] env[61907]: _type = "Task" [ 741.068479] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.076231] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243625, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.406578] env[61907]: INFO nova.compute.manager [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: 09d97063-d94f-439a-a811-b3039de1f4d7] Took 1.02 seconds to deallocate network for instance. [ 741.504808] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Releasing lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.505343] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 741.505552] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 741.508083] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1f32a498-c11d-4402-abfa-70bd3661044c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.523029] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc067f4d-c292-4994-a098-f4ef72c80100 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.532691] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.549106] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f517b2e4-1a83-4519-8a5b-65d489d16b6d could not be found. [ 741.549340] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 741.549569] env[61907]: INFO nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 741.549822] env[61907]: DEBUG oslo.service.loopingcall [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.553931] env[61907]: DEBUG nova.compute.manager [-] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 741.554068] env[61907]: DEBUG nova.network.neutron [-] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 741.574360] env[61907]: DEBUG nova.network.neutron [-] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 741.582181] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243625, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.599412] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.805553] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d217ef0-f434-4c91-9e74-547571b87976 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.813931] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca47e1e3-ee0e-4111-b1f6-dbc2e133664a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.843359] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0054c732-0c02-40ed-a7cb-b63b01ec8a10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.849640] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c04f150-fcb2-4829-ad7a-77c0b351a826 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.862674] env[61907]: DEBUG nova.compute.provider_tree [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.080193] env[61907]: DEBUG nova.network.neutron [-] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.081641] env[61907]: DEBUG oslo_vmware.api [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243625, 'name': PowerOnVM_Task, 'duration_secs': 0.968387} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.081641] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 742.081743] env[61907]: INFO nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Took 5.30 seconds to spawn the instance on the hypervisor. [ 742.082331] env[61907]: DEBUG nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 742.082921] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb698b63-1143-4b6f-bb80-6c19b807caea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.103872] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Releasing lock "refresh_cache-af575382-25f6-475a-8803-48714fa0a37c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.104832] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 742.104832] env[61907]: DEBUG nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 742.104832] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 742.125095] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.366079] env[61907]: DEBUG nova.scheduler.client.report [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 742.432710] env[61907]: INFO nova.scheduler.client.report [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Deleted allocations for instance 09d97063-d94f-439a-a811-b3039de1f4d7 [ 742.583830] env[61907]: INFO nova.compute.manager [-] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Took 1.03 seconds to deallocate network for instance. [ 742.586304] env[61907]: DEBUG nova.compute.claims [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 742.586500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.602025] env[61907]: INFO nova.compute.manager [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Took 26.16 seconds to build instance. [ 742.628017] env[61907]: DEBUG nova.network.neutron [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.870579] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.871224] env[61907]: ERROR nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Traceback (most recent call last): [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.driver.spawn(context, instance, image_meta, [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] vm_ref = self.build_virtual_machine(instance, [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.871224] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] for vif in network_info: [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return self._sync_wrapper(fn, *args, **kwargs) [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.wait() [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self[:] = self._gt.wait() [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return self._exit_event.wait() [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] current.throw(*self._exc) [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.871626] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] result = function(*args, **kwargs) [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] return func(*args, **kwargs) [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise e [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] nwinfo = self.network_api.allocate_for_instance( [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] created_port_ids = self._update_ports_for_instance( [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] with excutils.save_and_reraise_exception(): [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] self.force_reraise() [ 742.872280] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise self.value [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] updated_port = self._update_port( [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] _ensure_no_port_binding_failure(port) [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] raise exception.PortBindingFailed(port_id=port['id']) [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] nova.exception.PortBindingFailed: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. [ 742.872695] env[61907]: ERROR nova.compute.manager [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] [ 742.872695] env[61907]: DEBUG nova.compute.utils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 742.873423] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.940s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.874858] env[61907]: INFO nova.compute.claims [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.878196] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Build of instance a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3 was re-scheduled: Binding failed for port d0a29dc7-2683-4bc7-a386-8072507cc9d0, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 742.878625] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 742.878960] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.879155] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquired lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.879314] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 742.942243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d7a1f10c-29d8-47a8-9050-539462bcc33c tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "09d97063-d94f-439a-a811-b3039de1f4d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.460s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.104351] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89d2b7c1-cf5e-4d12-aedf-2cf068f73d50 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.699s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.131457] env[61907]: INFO nova.compute.manager [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] [instance: af575382-25f6-475a-8803-48714fa0a37c] Took 1.03 seconds to deallocate network for instance. [ 743.306200] env[61907]: INFO nova.compute.manager [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Rebuilding instance [ 743.347826] env[61907]: DEBUG nova.compute.manager [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 743.351491] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bc0190-50c7-4b48-af31-468d07fbf3e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.400897] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.443735] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 743.467300] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.607188] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 743.962168] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.971694] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Releasing lock "refresh_cache-a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.971921] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 743.972098] env[61907]: DEBUG nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 743.972268] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 743.988315] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.119359] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.119685] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.130685] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.156422] env[61907]: INFO nova.scheduler.client.report [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Deleted allocations for instance af575382-25f6-475a-8803-48714fa0a37c [ 744.201531] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59364bd0-00fd-46a1-940f-a3779ee32cc6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.209342] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0f7fbe-22a6-461b-aa2f-2cd8607d15c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.242425] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a037bb-edbd-4825-a0cd-45df22cf52d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.250056] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53520c16-b0c9-4f4b-867b-41788ece9bf0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.267243] env[61907]: DEBUG nova.compute.provider_tree [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.360896] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 744.361336] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4201686e-e4a2-4651-ada9-781e4576920b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.368089] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 744.368089] env[61907]: value = "task-1243626" [ 744.368089] env[61907]: _type = "Task" [ 744.368089] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.375993] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243626, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.492873] env[61907]: DEBUG nova.network.neutron [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.668208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-95ebc41f-0b16-4a95-9e75-93d88e15def5 tempest-MigrationsAdminTest-988735926 tempest-MigrationsAdminTest-988735926-project-member] Lock "af575382-25f6-475a-8803-48714fa0a37c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.730s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.770570] env[61907]: DEBUG nova.scheduler.client.report [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 744.877422] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243626, 'name': PowerOffVM_Task, 'duration_secs': 0.117011} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.877683] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 744.877925] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.878660] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b82f6ab-3c15-4630-b46a-66f48e15352b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.885017] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 744.885236] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e90f06a3-b881-4db8-b1cb-416fec6e4ae6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.912535] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 744.912777] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 744.912961] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Deleting the datastore file [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.913509] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5545542a-6914-4a20-8a2a-d79d1ff827c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.919294] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 744.919294] env[61907]: value = "task-1243628" [ 744.919294] env[61907]: _type = "Task" [ 744.919294] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.926457] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.995426] env[61907]: INFO nova.compute.manager [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3] Took 1.02 seconds to deallocate network for instance. [ 745.172033] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 745.275095] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.275661] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 745.281299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.516s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.429545] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094997} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.429846] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.430084] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 745.430296] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.695296] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.792052] env[61907]: DEBUG nova.compute.utils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.794732] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 745.794896] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.844555] env[61907]: DEBUG nova.policy [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8eb949aeb2a443eba91a011ef048cff0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9c2c0347510d4d7bb1f42056cc0bb8c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.038246] env[61907]: INFO nova.scheduler.client.report [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Deleted allocations for instance a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3 [ 746.127377] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aecf85f7-bb93-47f3-88a3-8f3467a31427 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.135481] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f2e2311-0c6d-4836-9448-2e74a49c6fd1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.168519] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Successfully created port: c9747636-55a8-49ff-ac51-d00435e9ba44 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 746.170862] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3387fa09-8995-444b-8979-3e7e7223cc0b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.178876] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd369fc-b89f-4ef5-8006-670d4233604c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.195133] env[61907]: DEBUG nova.compute.provider_tree [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.298726] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 746.474739] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.475008] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.475176] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.475362] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.475509] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.475654] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.475859] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.477638] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.477638] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.477638] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.477638] env[61907]: DEBUG nova.virt.hardware [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.478888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b45c1b-f92f-47d5-8666-f98193da9350 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.487791] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1949935-aa72-478d-b6f6-80c148013f93 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.506068] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.511944] env[61907]: DEBUG oslo.service.loopingcall [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.512240] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.512483] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-66c2873f-ca40-4d44-80de-c6e967b73f48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.529150] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.529150] env[61907]: value = "task-1243629" [ 746.529150] env[61907]: _type = "Task" [ 746.529150] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.538718] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243629, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.547723] env[61907]: DEBUG oslo_concurrency.lockutils [None req-094f7663-2702-4a0a-9dc8-94be109f7686 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "a1e088e5-5a9e-4c2c-ada1-b26fc7e730a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.855s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.698147] env[61907]: DEBUG nova.scheduler.client.report [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 747.036193] env[61907]: DEBUG nova.compute.manager [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Received event network-changed-c9747636-55a8-49ff-ac51-d00435e9ba44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 747.036441] env[61907]: DEBUG nova.compute.manager [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Refreshing instance network info cache due to event network-changed-c9747636-55a8-49ff-ac51-d00435e9ba44. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 747.036691] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] Acquiring lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.036862] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] Acquired lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.037063] env[61907]: DEBUG nova.network.neutron [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Refreshing network info cache for port c9747636-55a8-49ff-ac51-d00435e9ba44 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 747.045573] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243629, 'name': CreateVM_Task, 'duration_secs': 0.334487} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.046367] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 747.046827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.047482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.047482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 747.047614] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfd9edb8-b45b-41de-bd5e-cc658c1bf34f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.050039] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 747.055157] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 747.055157] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52acd9a0-9bd5-476c-dee5-a1904cb9f15c" [ 747.055157] env[61907]: _type = "Task" [ 747.055157] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.063507] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52acd9a0-9bd5-476c-dee5-a1904cb9f15c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.203233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.203878] env[61907]: ERROR nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Traceback (most recent call last): [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.driver.spawn(context, instance, image_meta, [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] vm_ref = self.build_virtual_machine(instance, [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.203878] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] for vif in network_info: [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return self._sync_wrapper(fn, *args, **kwargs) [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.wait() [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self[:] = self._gt.wait() [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return self._exit_event.wait() [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] current.throw(*self._exc) [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.205100] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] result = function(*args, **kwargs) [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] return func(*args, **kwargs) [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise e [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] nwinfo = self.network_api.allocate_for_instance( [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] created_port_ids = self._update_ports_for_instance( [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] with excutils.save_and_reraise_exception(): [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] self.force_reraise() [ 747.205479] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise self.value [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] updated_port = self._update_port( [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] _ensure_no_port_binding_failure(port) [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] raise exception.PortBindingFailed(port_id=port['id']) [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] nova.exception.PortBindingFailed: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. [ 747.205860] env[61907]: ERROR nova.compute.manager [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] [ 747.205860] env[61907]: DEBUG nova.compute.utils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.207117] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.878s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.207117] env[61907]: DEBUG nova.objects.instance [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lazy-loading 'resources' on Instance uuid 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.207517] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Build of instance 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827 was re-scheduled: Binding failed for port a6572b46-5a9b-464a-b414-90a03db283d7, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 747.207966] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 747.208210] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquiring lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.208356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Acquired lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.208517] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.240407] env[61907]: ERROR nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 747.240407] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.240407] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.240407] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.240407] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.240407] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.240407] env[61907]: ERROR nova.compute.manager raise self.value [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.240407] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.240407] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.240407] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.241263] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.241263] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.241263] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 747.241263] env[61907]: ERROR nova.compute.manager [ 747.241263] env[61907]: Traceback (most recent call last): [ 747.241263] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.241263] env[61907]: listener.cb(fileno) [ 747.241263] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.241263] env[61907]: result = function(*args, **kwargs) [ 747.241263] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.241263] env[61907]: return func(*args, **kwargs) [ 747.241263] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 747.241263] env[61907]: raise e [ 747.241263] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.241263] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 747.241263] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.241263] env[61907]: created_port_ids = self._update_ports_for_instance( [ 747.241263] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.241263] env[61907]: with excutils.save_and_reraise_exception(): [ 747.241263] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.241263] env[61907]: self.force_reraise() [ 747.241263] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.241263] env[61907]: raise self.value [ 747.241263] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.241263] env[61907]: updated_port = self._update_port( [ 747.241263] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.241263] env[61907]: _ensure_no_port_binding_failure(port) [ 747.241263] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.241263] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.242028] env[61907]: nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 747.242028] env[61907]: Removing descriptor: 15 [ 747.309153] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 747.338391] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.338660] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.338810] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.338990] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.339157] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.339307] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.339512] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.339677] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.339846] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.339974] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.340193] env[61907]: DEBUG nova.virt.hardware [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.342640] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36939ee-a5be-45c4-a454-560cc2542af0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.352203] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329bf93b-0f5d-47eb-9760-16d741ea6427 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.368019] env[61907]: ERROR nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Traceback (most recent call last): [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] yield resources [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.driver.spawn(context, instance, image_meta, [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] vm_ref = self.build_virtual_machine(instance, [ 747.368019] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] for vif in network_info: [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return self._sync_wrapper(fn, *args, **kwargs) [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.wait() [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self[:] = self._gt.wait() [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return self._exit_event.wait() [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.368448] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] current.throw(*self._exc) [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] result = function(*args, **kwargs) [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return func(*args, **kwargs) [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise e [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] nwinfo = self.network_api.allocate_for_instance( [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] created_port_ids = self._update_ports_for_instance( [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] with excutils.save_and_reraise_exception(): [ 747.368799] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.force_reraise() [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise self.value [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] updated_port = self._update_port( [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] _ensure_no_port_binding_failure(port) [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise exception.PortBindingFailed(port_id=port['id']) [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 747.369228] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] [ 747.369228] env[61907]: INFO nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Terminating instance [ 747.556770] env[61907]: DEBUG nova.network.neutron [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.572206] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52acd9a0-9bd5-476c-dee5-a1904cb9f15c, 'name': SearchDatastore_Task, 'duration_secs': 0.00847} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.572686] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.573060] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 747.573409] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.573672] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.574467] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 747.574467] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f3da91d-c463-4436-972c-5fb702d4a64e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.582507] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 747.583158] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 747.583982] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8229c349-685d-4a09-9b28-0b5a3cba4fe4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.590785] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.590785] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 747.590785] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523543e0-abdb-887a-a34c-befd83a1d9d2" [ 747.590785] env[61907]: _type = "Task" [ 747.590785] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.600642] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523543e0-abdb-887a-a34c-befd83a1d9d2, 'name': SearchDatastore_Task, 'duration_secs': 0.007903} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.601400] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0f8962b-aa5d-4456-87a8-ff217339da33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.606042] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 747.606042] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5248daf8-3151-28c8-4f92-eae9f79477fb" [ 747.606042] env[61907]: _type = "Task" [ 747.606042] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.613464] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5248daf8-3151-28c8-4f92-eae9f79477fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.659707] env[61907]: DEBUG nova.network.neutron [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.728363] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.840133] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.874741] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquiring lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.025117] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594c4cb1-20c6-45c6-af43-26176488110b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.033344] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525ddbf9-00f0-48c5-98b7-37673a7dd12b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.068404] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c41b3a4f-e86a-4ad9-ae68-35af6cf104d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.073911] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0584ac60-70fd-4402-b0d1-329d75b1f5bf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.086819] env[61907]: DEBUG nova.compute.provider_tree [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.115665] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5248daf8-3151-28c8-4f92-eae9f79477fb, 'name': SearchDatastore_Task, 'duration_secs': 0.008728} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.115906] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.116178] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 748.116513] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e05d68ca-0f86-486f-bc53-42b53ef019ca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.122477] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 748.122477] env[61907]: value = "task-1243630" [ 748.122477] env[61907]: _type = "Task" [ 748.122477] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.132089] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243630, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.162322] env[61907]: DEBUG oslo_concurrency.lockutils [req-02c5596c-32fb-4d3a-b6c8-5e5fa8b8c1aa req-a5670ca3-ec52-460c-b916-ef805a8ace9b service nova] Releasing lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.163084] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquired lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.163286] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 748.342256] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Releasing lock "refresh_cache-4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.342256] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 748.342589] env[61907]: DEBUG nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 748.342700] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.378859] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.592192] env[61907]: DEBUG nova.scheduler.client.report [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 748.632258] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243630, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468433} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.632570] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 748.632834] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.633169] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ca8581ce-73a1-4f3d-a1f7-ca6e7fe581c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.639048] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 748.639048] env[61907]: value = "task-1243631" [ 748.639048] env[61907]: _type = "Task" [ 748.639048] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.646844] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243631, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.690807] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.772866] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.879849] env[61907]: DEBUG nova.network.neutron [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.060951] env[61907]: DEBUG nova.compute.manager [req-c7d2676a-7a9d-4cfa-a8a2-00298e9fd54b req-27c3ded4-fc8d-4969-b09e-aa2d921eafca service nova] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Received event network-vif-deleted-c9747636-55a8-49ff-ac51-d00435e9ba44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 749.095544] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.890s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.097848] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.398s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.120715] env[61907]: INFO nova.scheduler.client.report [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Deleted allocations for instance 67a7aef0-ed99-4bde-b2f0-33c36764a5d7 [ 749.150611] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243631, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057622} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.150611] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.151395] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c360ffe-ed6e-4cac-9b93-cb7cdb909cb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.170278] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.170879] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90c4574c-a3d2-49e7-8a9d-f88f70ee55fd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.191248] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 749.191248] env[61907]: value = "task-1243632" [ 749.191248] env[61907]: _type = "Task" [ 749.191248] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.200112] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243632, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.277266] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Releasing lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.277266] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 749.277266] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 749.277555] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14134635-4108-4299-a2a0-e82fcd7d35a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.287683] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37304fc-4285-4ab6-aeec-27cceb03d848 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.312287] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d could not be found. [ 749.312533] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 749.312738] env[61907]: INFO nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 749.312992] env[61907]: DEBUG oslo.service.loopingcall [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.313721] env[61907]: DEBUG nova.compute.manager [-] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 749.313721] env[61907]: DEBUG nova.network.neutron [-] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 749.332983] env[61907]: DEBUG nova.network.neutron [-] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 749.386063] env[61907]: INFO nova.compute.manager [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] [instance: 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827] Took 1.04 seconds to deallocate network for instance. [ 749.629390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad07f441-2ea4-43bf-84f0-5a5aff913288 tempest-ServersAaction247Test-547639548 tempest-ServersAaction247Test-547639548-project-member] Lock "67a7aef0-ed99-4bde-b2f0-33c36764a5d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.492s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.703737] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243632, 'name': ReconfigVM_Task, 'duration_secs': 0.297017} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.704106] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfigured VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.704923] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0831aed-3481-49d9-9a05-217d0731f0dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.713933] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 749.713933] env[61907]: value = "task-1243633" [ 749.713933] env[61907]: _type = "Task" [ 749.713933] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.724030] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243633, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.835776] env[61907]: DEBUG nova.network.neutron [-] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.944888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bbc2df-8d09-475f-8a24-4daeb16dcef5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.953267] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b038100-5933-4541-93f8-a3cbb23e4c7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.986062] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7105634f-77dc-487a-b2d1-8460c89eb233 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.994316] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f296795-c936-4453-95a3-94df5105eedd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.015944] env[61907]: DEBUG nova.compute.provider_tree [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.226282] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243633, 'name': Rename_Task, 'duration_secs': 0.125384} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.226282] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 750.226282] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1617745-0a85-406c-bc7b-609d65f9611b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.232746] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 750.232746] env[61907]: value = "task-1243634" [ 750.232746] env[61907]: _type = "Task" [ 750.232746] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.240523] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243634, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.338314] env[61907]: INFO nova.compute.manager [-] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Took 1.02 seconds to deallocate network for instance. [ 750.340625] env[61907]: DEBUG nova.compute.claims [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 750.340976] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.422562] env[61907]: INFO nova.scheduler.client.report [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Deleted allocations for instance 4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827 [ 750.520689] env[61907]: DEBUG nova.scheduler.client.report [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 750.743428] env[61907]: DEBUG oslo_vmware.api [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243634, 'name': PowerOnVM_Task, 'duration_secs': 0.404356} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.745960] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 750.745960] env[61907]: DEBUG nova.compute.manager [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 750.745960] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435d4b01-a011-4bbb-a050-92f2d985ca69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.809772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.809772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.933398] env[61907]: DEBUG oslo_concurrency.lockutils [None req-04b2e788-152c-4dd7-81c3-fbbca0720025 tempest-ServersV294TestFqdnHostnames-2020507215 tempest-ServersV294TestFqdnHostnames-2020507215-project-member] Lock "4f86d6b8-5fb9-4a20-ad49-a3d7ceb3e827" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.747s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.026994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.027581] env[61907]: ERROR nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Traceback (most recent call last): [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.driver.spawn(context, instance, image_meta, [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] vm_ref = self.build_virtual_machine(instance, [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.027581] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] for vif in network_info: [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self._sync_wrapper(fn, *args, **kwargs) [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.wait() [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self[:] = self._gt.wait() [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self._exit_event.wait() [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] result = hub.switch() [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.028119] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return self.greenlet.switch() [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] result = function(*args, **kwargs) [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] return func(*args, **kwargs) [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise e [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] nwinfo = self.network_api.allocate_for_instance( [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] created_port_ids = self._update_ports_for_instance( [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] with excutils.save_and_reraise_exception(): [ 751.028493] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] self.force_reraise() [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise self.value [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] updated_port = self._update_port( [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] _ensure_no_port_binding_failure(port) [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] raise exception.PortBindingFailed(port_id=port['id']) [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] nova.exception.PortBindingFailed: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. [ 751.028848] env[61907]: ERROR nova.compute.manager [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] [ 751.029214] env[61907]: DEBUG nova.compute.utils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.029960] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.862s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.031542] env[61907]: INFO nova.compute.claims [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.035344] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Build of instance ea6d2b19-be25-4da3-b8af-11c3dca04df8 was re-scheduled: Binding failed for port fe20272c-cdd7-495c-818b-12ecb67f7ac2, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 751.038930] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 751.039173] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquiring lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.039338] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Acquired lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.039520] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.259759] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.437066] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 751.561208] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.634382] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.963406] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.137184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Releasing lock "refresh_cache-ea6d2b19-be25-4da3-b8af-11c3dca04df8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.137374] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 752.137538] env[61907]: DEBUG nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 752.137722] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.177080] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.370779] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba7938e-5f53-4220-ac54-da6065c9904c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.378270] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c157b77d-fa56-431b-9934-2b0816668717 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.408502] env[61907]: INFO nova.compute.manager [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Rebuilding instance [ 752.411555] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5012f110-66a0-477a-8532-076fde88788c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.418616] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374c2b1b-affa-462a-9cd3-0bdf92de8b67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.435531] env[61907]: DEBUG nova.compute.provider_tree [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.463945] env[61907]: DEBUG nova.compute.manager [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 752.465064] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4416cf28-36b6-4835-96bd-264e35cd1fdf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.682809] env[61907]: DEBUG nova.network.neutron [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.940298] env[61907]: DEBUG nova.scheduler.client.report [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 753.188241] env[61907]: INFO nova.compute.manager [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] [instance: ea6d2b19-be25-4da3-b8af-11c3dca04df8] Took 1.05 seconds to deallocate network for instance. [ 753.445623] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.446205] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 753.449133] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.360s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.450702] env[61907]: INFO nova.compute.claims [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.479681] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 753.479984] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c86551fa-64f6-483e-a6c4-eebd310f51d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.491700] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 753.491700] env[61907]: value = "task-1243635" [ 753.491700] env[61907]: _type = "Task" [ 753.491700] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.501837] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.961337] env[61907]: DEBUG nova.compute.utils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 753.966123] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 753.966123] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 754.002227] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243635, 'name': PowerOffVM_Task, 'duration_secs': 0.18402} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.002550] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 754.002738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 754.003500] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dec419a-9ed5-489e-9679-cada2a73f9f3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.007968] env[61907]: DEBUG nova.policy [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb8984d341bb4fdf892048b6c9ec0be3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'be91ad381ac848bfb21a4673d9fc8e77', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 754.014116] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 754.014116] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6b3432a-6ad8-4780-b6fd-ca4127e23b0c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.038306] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 754.038523] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 754.038704] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Deleting the datastore file [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.038969] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e1c5fa3-9ddd-4d2c-a496-f35edf9ce029 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.044753] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 754.044753] env[61907]: value = "task-1243637" [ 754.044753] env[61907]: _type = "Task" [ 754.044753] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.052600] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.225441] env[61907]: INFO nova.scheduler.client.report [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Deleted allocations for instance ea6d2b19-be25-4da3-b8af-11c3dca04df8 [ 754.344741] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Successfully created port: 30ec1c1b-b2e0-4890-a573-18042eb14e44 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.467584] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 754.559134] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.088248} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.559715] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.560106] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 754.560691] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.642597] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Successfully created port: 8d52ecc3-f7f2-4a86-8447-f25df78b687f {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 754.735028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7c680ad-6a2b-42c1-b07c-e4147986c758 tempest-ListImageFiltersTestJSON-1341127333 tempest-ListImageFiltersTestJSON-1341127333-project-member] Lock "ea6d2b19-be25-4da3-b8af-11c3dca04df8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.959s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.781635] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32af29d3-3fbc-43e9-b4c1-fa800fe340b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.790147] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf450eb9-5062-4a84-b549-e2a59f45d69d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.823762] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd7cd95-a120-4750-9a75-307f032a0e0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.830787] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c45cfb6-51c4-4797-b8a9-55000180404e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.848513] env[61907]: DEBUG nova.compute.provider_tree [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.237203] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 755.353868] env[61907]: DEBUG nova.scheduler.client.report [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 755.481707] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 755.516879] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.517134] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.517294] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.517476] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.517618] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.520021] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.520021] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.520021] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.520021] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.520021] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.520397] env[61907]: DEBUG nova.virt.hardware [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.520397] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b06c1c2-aab8-409f-9440-ded29f0f279d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.529000] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e82df4-0f7b-48ae-ae8f-47b102d08965 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.535062] env[61907]: DEBUG nova.compute.manager [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Received event network-changed-30ec1c1b-b2e0-4890-a573-18042eb14e44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 755.535318] env[61907]: DEBUG nova.compute.manager [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Refreshing instance network info cache due to event network-changed-30ec1c1b-b2e0-4890-a573-18042eb14e44. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 755.535608] env[61907]: DEBUG oslo_concurrency.lockutils [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] Acquiring lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.535816] env[61907]: DEBUG oslo_concurrency.lockutils [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] Acquired lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.536038] env[61907]: DEBUG nova.network.neutron [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Refreshing network info cache for port 30ec1c1b-b2e0-4890-a573-18042eb14e44 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.595904] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 755.596204] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 755.596341] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 755.596525] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 755.596682] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 755.596866] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 755.597541] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 755.597809] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 755.598062] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 755.598301] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 755.598695] env[61907]: DEBUG nova.virt.hardware [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.599697] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fead5acc-4598-4a65-bf51-267129bb50bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.609730] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451b9ea8-f83d-4ade-a1c7-81e249bcaae8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.623928] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 755.630920] env[61907]: DEBUG oslo.service.loopingcall [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.630920] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 755.630920] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c3a5f724-db7c-466c-9272-7ae3bf78a41b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.647834] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 755.647834] env[61907]: value = "task-1243638" [ 755.647834] env[61907]: _type = "Task" [ 755.647834] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.657347] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243638, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.772409] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.808513] env[61907]: ERROR nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 755.808513] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 755.808513] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.808513] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.808513] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.808513] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.808513] env[61907]: ERROR nova.compute.manager raise self.value [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.808513] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 755.808513] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.808513] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 755.809054] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.809054] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 755.809054] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 755.809054] env[61907]: ERROR nova.compute.manager [ 755.809054] env[61907]: Traceback (most recent call last): [ 755.809054] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 755.809054] env[61907]: listener.cb(fileno) [ 755.809054] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.809054] env[61907]: result = function(*args, **kwargs) [ 755.809054] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.809054] env[61907]: return func(*args, **kwargs) [ 755.809054] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 755.809054] env[61907]: raise e [ 755.809054] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 755.809054] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 755.809054] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.809054] env[61907]: created_port_ids = self._update_ports_for_instance( [ 755.809054] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.809054] env[61907]: with excutils.save_and_reraise_exception(): [ 755.809054] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.809054] env[61907]: self.force_reraise() [ 755.809054] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.809054] env[61907]: raise self.value [ 755.809054] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.809054] env[61907]: updated_port = self._update_port( [ 755.809054] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.809054] env[61907]: _ensure_no_port_binding_failure(port) [ 755.809054] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.809054] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 755.809940] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 755.809940] env[61907]: Removing descriptor: 15 [ 755.809940] env[61907]: ERROR nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Traceback (most recent call last): [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] yield resources [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.driver.spawn(context, instance, image_meta, [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 755.809940] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] vm_ref = self.build_virtual_machine(instance, [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] vif_infos = vmwarevif.get_vif_info(self._session, [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] for vif in network_info: [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self._sync_wrapper(fn, *args, **kwargs) [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.wait() [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self[:] = self._gt.wait() [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self._exit_event.wait() [ 755.810339] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] result = hub.switch() [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self.greenlet.switch() [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] result = function(*args, **kwargs) [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return func(*args, **kwargs) [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise e [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] nwinfo = self.network_api.allocate_for_instance( [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 755.810740] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] created_port_ids = self._update_ports_for_instance( [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] with excutils.save_and_reraise_exception(): [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.force_reraise() [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise self.value [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] updated_port = self._update_port( [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] _ensure_no_port_binding_failure(port) [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 755.811159] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise exception.PortBindingFailed(port_id=port['id']) [ 755.811529] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 755.811529] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] [ 755.811529] env[61907]: INFO nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Terminating instance [ 755.862684] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.863270] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 755.867429] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.082s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.868552] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.868552] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 755.868948] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.282s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.873138] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931d4075-1fd0-49c0-8cb5-899a03ecbd03 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.884814] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f919af08-c231-44c3-8c93-6e35ff4a9af0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.900328] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a59b6da-6147-44f9-9768-712360778898 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.906509] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d08d22-14b0-4ee3-a6b6-8fa0fb5c5939 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.936799] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181534MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 755.937679] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.089142] env[61907]: DEBUG nova.network.neutron [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.161715] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243638, 'name': CreateVM_Task, 'duration_secs': 0.294936} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.161715] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 756.161715] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.161905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.162345] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 756.162711] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55ec4969-cc4e-4348-880f-0e0d4843ecf1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.168296] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 756.168296] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5bdac-f872-be65-b631-dfdf9a5001a3" [ 756.168296] env[61907]: _type = "Task" [ 756.168296] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.179968] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5bdac-f872-be65-b631-dfdf9a5001a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.233787] env[61907]: DEBUG nova.network.neutron [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.315038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.369410] env[61907]: DEBUG nova.compute.utils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.371737] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 756.372618] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 756.463268] env[61907]: DEBUG nova.policy [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1689ce713672416789830dcf837a5e5c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c6d7c24362f042ef830de175a2939851', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 756.681814] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5bdac-f872-be65-b631-dfdf9a5001a3, 'name': SearchDatastore_Task, 'duration_secs': 0.011765} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.682151] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.682413] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 756.682605] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.682751] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.682929] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 756.683207] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d895e3ef-7393-4040-b23b-6f05c8c87b47 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.690792] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 756.690965] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 756.691730] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-922513bc-c8ca-4305-a70d-4a9fb1cffd24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.703023] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 756.703023] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52df6d28-0232-cdbb-ca3c-9c2eef813461" [ 756.703023] env[61907]: _type = "Task" [ 756.703023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.710916] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52df6d28-0232-cdbb-ca3c-9c2eef813461, 'name': SearchDatastore_Task, 'duration_secs': 0.007886} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.714204] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-961e1f40-9162-4901-80ba-497fd7b3276b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.723018] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 756.723018] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b636ed-f7b2-7540-f839-08b69ea736eb" [ 756.723018] env[61907]: _type = "Task" [ 756.723018] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.729513] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b636ed-f7b2-7540-f839-08b69ea736eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.736242] env[61907]: DEBUG oslo_concurrency.lockutils [req-91015284-09fa-47d7-809d-6f8c7ff56a63 req-280d5c28-ad9d-43bd-9b0a-81444d589f99 service nova] Releasing lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.736817] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquired lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.737056] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.743355] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace95448-28c3-46fe-b55a-ad9b35002d39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.750352] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b55b6d-2f26-4a4a-b694-62b594897028 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.784722] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b209a78-923a-4b22-8c29-df166c6f248d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.792259] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e853311-d401-4875-b9bd-d619f20df3f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.805647] env[61907]: DEBUG nova.compute.provider_tree [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.871323] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.872122] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.878677] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 756.994736] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Successfully created port: 362874eb-244f-4fa8-a357-bde8871e3075 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.236073] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b636ed-f7b2-7540-f839-08b69ea736eb, 'name': SearchDatastore_Task, 'duration_secs': 0.007925} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.238642] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.238848] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 757.241043] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3c22f715-6343-4fb6-896b-9c85647d2bb7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.249107] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 757.249107] env[61907]: value = "task-1243639" [ 757.249107] env[61907]: _type = "Task" [ 757.249107] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.260017] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243639, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.291659] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.313110] env[61907]: DEBUG nova.scheduler.client.report [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 757.467450] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.640576] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "23526967-cd8f-4581-b9c5-1c270d385163" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.640880] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.741915] env[61907]: DEBUG nova.compute.manager [req-746ff95f-b5b9-478c-a994-ff5fd864420d req-3ddd553d-6e17-47a6-bf9b-4927383213f6 service nova] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Received event network-vif-deleted-30ec1c1b-b2e0-4890-a573-18042eb14e44 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 757.758956] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243639, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.482187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.759280] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 757.759509] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.759854] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf497842-9776-45b8-9118-40778b116277 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.767093] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 757.767093] env[61907]: value = "task-1243640" [ 757.767093] env[61907]: _type = "Task" [ 757.767093] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.776920] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243640, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.821234] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.821872] env[61907]: ERROR nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Traceback (most recent call last): [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.driver.spawn(context, instance, image_meta, [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] vm_ref = self.build_virtual_machine(instance, [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] vif_infos = vmwarevif.get_vif_info(self._session, [ 757.821872] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] for vif in network_info: [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return self._sync_wrapper(fn, *args, **kwargs) [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.wait() [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self[:] = self._gt.wait() [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return self._exit_event.wait() [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] current.throw(*self._exc) [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 757.822344] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] result = function(*args, **kwargs) [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] return func(*args, **kwargs) [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise e [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] nwinfo = self.network_api.allocate_for_instance( [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] created_port_ids = self._update_ports_for_instance( [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] with excutils.save_and_reraise_exception(): [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] self.force_reraise() [ 757.822715] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise self.value [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] updated_port = self._update_port( [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] _ensure_no_port_binding_failure(port) [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] raise exception.PortBindingFailed(port_id=port['id']) [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] nova.exception.PortBindingFailed: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. [ 757.823095] env[61907]: ERROR nova.compute.manager [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] [ 757.823095] env[61907]: DEBUG nova.compute.utils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 757.824256] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.862s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.825450] env[61907]: INFO nova.compute.claims [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.828765] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Build of instance f517b2e4-1a83-4519-8a5b-65d489d16b6d was re-scheduled: Binding failed for port 38a00017-9f7b-4899-8e16-3e432ebb3cf1, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 757.829230] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 757.829464] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquiring lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.829614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Acquired lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.829774] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.894794] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 757.926883] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.931155] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.931561] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.931824] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.931981] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.932148] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.932441] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.932619] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.932798] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.932965] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.933156] env[61907]: DEBUG nova.virt.hardware [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.934332] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7894e2-90c0-4e39-8fc2-abae13c39bab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.945583] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82b59c8-f7e0-4a85-8164-f77f37590501 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.969795] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Releasing lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.970268] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 757.970504] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.970941] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d32e826e-80e3-412e-903b-2362a841a3be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.979375] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a27d7ac-ba9d-466a-84fb-4ac2cc850595 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.001408] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 215d0f35-5d19-4db4-996d-571e2970e19c could not be found. [ 758.001408] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 758.001857] env[61907]: INFO nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 758.001857] env[61907]: DEBUG oslo.service.loopingcall [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 758.002093] env[61907]: DEBUG nova.compute.manager [-] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 758.002204] env[61907]: DEBUG nova.network.neutron [-] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.060614] env[61907]: DEBUG nova.network.neutron [-] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.278404] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243640, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066001} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.279177] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.282140] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d663d1-c0cf-4bba-92f6-fdce346f5cdb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.307327] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.307841] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2caf5684-0bb7-4dd0-975a-efb5acb5c2b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.330274] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 758.330274] env[61907]: value = "task-1243641" [ 758.330274] env[61907]: _type = "Task" [ 758.330274] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.343461] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243641, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.360138] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.442298] env[61907]: ERROR nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 758.442298] env[61907]: ERROR nova.compute.manager Traceback (most recent call last): [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.442298] env[61907]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.442298] env[61907]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.442298] env[61907]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.442298] env[61907]: ERROR nova.compute.manager self.force_reraise() [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.442298] env[61907]: ERROR nova.compute.manager raise self.value [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.442298] env[61907]: ERROR nova.compute.manager updated_port = self._update_port( [ 758.442298] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.442298] env[61907]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 758.442967] env[61907]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.442967] env[61907]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 758.442967] env[61907]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 758.442967] env[61907]: ERROR nova.compute.manager [ 758.442967] env[61907]: Traceback (most recent call last): [ 758.442967] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 758.442967] env[61907]: listener.cb(fileno) [ 758.442967] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.442967] env[61907]: result = function(*args, **kwargs) [ 758.442967] env[61907]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.442967] env[61907]: return func(*args, **kwargs) [ 758.442967] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 758.442967] env[61907]: raise e [ 758.442967] env[61907]: File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.442967] env[61907]: nwinfo = self.network_api.allocate_for_instance( [ 758.442967] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.442967] env[61907]: created_port_ids = self._update_ports_for_instance( [ 758.442967] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.442967] env[61907]: with excutils.save_and_reraise_exception(): [ 758.442967] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.442967] env[61907]: self.force_reraise() [ 758.442967] env[61907]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.442967] env[61907]: raise self.value [ 758.442967] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.442967] env[61907]: updated_port = self._update_port( [ 758.442967] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.442967] env[61907]: _ensure_no_port_binding_failure(port) [ 758.442967] env[61907]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.442967] env[61907]: raise exception.PortBindingFailed(port_id=port['id']) [ 758.443760] env[61907]: nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 758.443760] env[61907]: Removing descriptor: 15 [ 758.443760] env[61907]: ERROR nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Traceback (most recent call last): [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 2900, in _build_resources [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] yield resources [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.driver.spawn(context, instance, image_meta, [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self._vmops.spawn(context, instance, image_meta, injected_files, [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 758.443760] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] vm_ref = self.build_virtual_machine(instance, [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] vif_infos = vmwarevif.get_vif_info(self._session, [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] for vif in network_info: [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self._sync_wrapper(fn, *args, **kwargs) [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.wait() [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self[:] = self._gt.wait() [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self._exit_event.wait() [ 758.444190] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] result = hub.switch() [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self.greenlet.switch() [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] result = function(*args, **kwargs) [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return func(*args, **kwargs) [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise e [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] nwinfo = self.network_api.allocate_for_instance( [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 758.444943] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] created_port_ids = self._update_ports_for_instance( [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] with excutils.save_and_reraise_exception(): [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.force_reraise() [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise self.value [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] updated_port = self._update_port( [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] _ensure_no_port_binding_failure(port) [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 758.446957] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise exception.PortBindingFailed(port_id=port['id']) [ 758.447384] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 758.447384] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] [ 758.447384] env[61907]: INFO nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Terminating instance [ 758.518097] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.839544] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243641, 'name': ReconfigVM_Task, 'duration_secs': 0.294355} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.840293] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Reconfigured VM instance instance-0000002b to attach disk [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314/da7ddc92-f3e3-4b4d-a14b-13fd25a72314.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 758.840293] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c1006aa1-92e7-4a11-ba32-fe3d9f0c0f47 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.848334] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 758.848334] env[61907]: value = "task-1243642" [ 758.848334] env[61907]: _type = "Task" [ 758.848334] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.856229] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243642, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.925809] env[61907]: DEBUG nova.network.neutron [-] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.950140] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquiring lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.950419] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquired lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.950664] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.020655] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Releasing lock "refresh_cache-f517b2e4-1a83-4519-8a5b-65d489d16b6d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.020889] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 759.021952] env[61907]: DEBUG nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 759.021952] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 759.037769] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.161221] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b4e95c-e9b8-4665-849e-fc20ddd40b22 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.169168] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67be2c9e-9fb6-49d3-a86f-5150a4bd98c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.204785] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be89c0f5-344c-44e2-a974-fe84e156c35a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.212727] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b726450-5fd0-40be-84af-545b70961bc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.228029] env[61907]: DEBUG nova.compute.provider_tree [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 759.357898] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243642, 'name': Rename_Task, 'duration_secs': 0.134854} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.358184] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 759.358474] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2e1e115-1b16-4ef9-aac6-9427e6e884a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.365145] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Waiting for the task: (returnval){ [ 759.365145] env[61907]: value = "task-1243643" [ 759.365145] env[61907]: _type = "Task" [ 759.365145] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.373410] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243643, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.429343] env[61907]: INFO nova.compute.manager [-] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Took 1.43 seconds to deallocate network for instance. [ 759.431747] env[61907]: DEBUG nova.compute.claims [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 759.431747] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.473228] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.541203] env[61907]: DEBUG nova.network.neutron [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.589568] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.733616] env[61907]: DEBUG nova.scheduler.client.report [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 759.877140] env[61907]: DEBUG nova.compute.manager [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Received event network-changed-362874eb-244f-4fa8-a357-bde8871e3075 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 759.877458] env[61907]: DEBUG nova.compute.manager [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Refreshing instance network info cache due to event network-changed-362874eb-244f-4fa8-a357-bde8871e3075. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 759.877548] env[61907]: DEBUG oslo_concurrency.lockutils [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] Acquiring lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.886556] env[61907]: DEBUG oslo_vmware.api [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Task: {'id': task-1243643, 'name': PowerOnVM_Task, 'duration_secs': 0.413901} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.887553] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.887553] env[61907]: DEBUG nova.compute.manager [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 759.887998] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ead523-5890-43b3-8307-106acf821e7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.045278] env[61907]: INFO nova.compute.manager [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] [instance: f517b2e4-1a83-4519-8a5b-65d489d16b6d] Took 1.02 seconds to deallocate network for instance. [ 760.091957] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Releasing lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.092407] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 760.092603] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.093332] env[61907]: DEBUG oslo_concurrency.lockutils [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] Acquired lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.093332] env[61907]: DEBUG nova.network.neutron [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Refreshing network info cache for port 362874eb-244f-4fa8-a357-bde8871e3075 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.094571] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68e9b8cd-92d6-4412-bf42-4a5f166fba56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.104171] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54a58a6-7e38-4e08-b802-04efd8af6dc1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.127389] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 45c58b31-d0cc-4b36-b744-448a38e11266 could not be found. [ 760.127626] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.127806] env[61907]: INFO nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Took 0.04 seconds to destroy the instance on the hypervisor. [ 760.128081] env[61907]: DEBUG oslo.service.loopingcall [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.128351] env[61907]: DEBUG nova.compute.manager [-] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 760.128440] env[61907]: DEBUG nova.network.neutron [-] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.151267] env[61907]: DEBUG nova.network.neutron [-] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.241726] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.242282] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 760.245028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.115s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.246492] env[61907]: INFO nova.compute.claims [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 760.404370] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.615129] env[61907]: DEBUG nova.network.neutron [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.653633] env[61907]: DEBUG nova.network.neutron [-] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.691670] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.691902] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.692131] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.692341] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.692526] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.694770] env[61907]: DEBUG nova.network.neutron [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.696531] env[61907]: INFO nova.compute.manager [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Terminating instance [ 760.751304] env[61907]: DEBUG nova.compute.utils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 760.760987] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 760.763286] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 760.763451] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.801967] env[61907]: DEBUG nova.policy [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '772dbe17f5cd4dc9904197e708e3f316', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3aab40034a4e4ccdb169abfc11f3b02f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.953819] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "da490943-511b-4776-8f16-4f51c3b055c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.954137] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.079295] env[61907]: INFO nova.scheduler.client.report [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Deleted allocations for instance f517b2e4-1a83-4519-8a5b-65d489d16b6d [ 761.129251] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Successfully created port: 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 761.156159] env[61907]: INFO nova.compute.manager [-] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Took 1.03 seconds to deallocate network for instance. [ 761.158607] env[61907]: DEBUG nova.compute.claims [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Aborting claim: {{(pid=61907) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.158780] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.198296] env[61907]: DEBUG oslo_concurrency.lockutils [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] Releasing lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.198296] env[61907]: DEBUG nova.compute.manager [req-8e4d2a1d-3c5a-4d58-a10c-5f7dd9575d3d req-cd5c63bf-6478-4336-8e38-5950a894c458 service nova] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Received event network-vif-deleted-362874eb-244f-4fa8-a357-bde8871e3075 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 761.200387] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "refresh_cache-da7ddc92-f3e3-4b4d-a14b-13fd25a72314" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.200387] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquired lock "refresh_cache-da7ddc92-f3e3-4b4d-a14b-13fd25a72314" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.200387] env[61907]: DEBUG nova.network.neutron [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.581861] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035a57d3-fab2-4b0f-a915-4be8b9986add {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.589309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1c557fb0-ce55-4a8c-89fb-89876f44dc06 tempest-ServerActionsTestOtherA-1647940953 tempest-ServerActionsTestOtherA-1647940953-project-member] Lock "f517b2e4-1a83-4519-8a5b-65d489d16b6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.649s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.591385] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d5cf68-f775-4a78-a746-acd17be73fa4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.626859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6b4a1b-d8d9-41af-8d34-f6dbaa1b9b03 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.634704] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b6efc3-af04-4e4b-a611-64be824d39e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.649466] env[61907]: DEBUG nova.compute.provider_tree [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 761.724881] env[61907]: DEBUG nova.network.neutron [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.776695] env[61907]: DEBUG nova.network.neutron [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.779042] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 761.802948] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.803199] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.803354] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.803533] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.803676] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.803819] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.804025] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.804186] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.804347] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.804596] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.804798] env[61907]: DEBUG nova.virt.hardware [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.805870] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa62838-1fe9-4cee-af5e-05ec9959c71f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.814017] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34596490-be64-4b2e-bef6-ef1f03e6fe6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.099573] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 762.154089] env[61907]: DEBUG nova.scheduler.client.report [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 762.283610] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Releasing lock "refresh_cache-da7ddc92-f3e3-4b4d-a14b-13fd25a72314" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.283610] env[61907]: DEBUG nova.compute.manager [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 762.283610] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.283610] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddd796e-9fae-4c3a-a6aa-e58a9a2a644a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.293019] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 762.293019] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-530a3691-c720-41dd-abb4-8bf8aee34460 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.299254] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 762.299254] env[61907]: value = "task-1243644" [ 762.299254] env[61907]: _type = "Task" [ 762.299254] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.307529] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243644, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.633244] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.657303] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.657848] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 762.664022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.966s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.664022] env[61907]: INFO nova.compute.claims [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.760073] env[61907]: DEBUG nova.compute.manager [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 762.760306] env[61907]: DEBUG oslo_concurrency.lockutils [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.760701] env[61907]: DEBUG oslo_concurrency.lockutils [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.760983] env[61907]: DEBUG oslo_concurrency.lockutils [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.761102] env[61907]: DEBUG nova.compute.manager [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] No waiting events found dispatching network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 762.761259] env[61907]: WARNING nova.compute.manager [req-05295848-b7b0-47c0-a754-b6a99d0fabd3 req-ae8ffbc0-15d5-412d-b864-d554e0cab0ff service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received unexpected event network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 for instance with vm_state building and task_state spawning. [ 762.811949] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243644, 'name': PowerOffVM_Task, 'duration_secs': 0.113321} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.811949] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 762.811949] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 762.811949] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-048c689b-5c13-4cea-8a48-43e07bdc310e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.833788] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 762.833998] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 762.834191] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Deleting the datastore file [datastore2] da7ddc92-f3e3-4b4d-a14b-13fd25a72314 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 762.834447] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1c29e6d-e8cb-40e3-bdb5-74f9df46f161 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.851827] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for the task: (returnval){ [ 762.851827] env[61907]: value = "task-1243646" [ 762.851827] env[61907]: _type = "Task" [ 762.851827] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.859723] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243646, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.937442] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Successfully updated port: 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 763.166430] env[61907]: DEBUG nova.compute.utils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 763.167926] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 763.168112] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 763.249815] env[61907]: DEBUG nova.policy [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c1ba3b490b60492d9117e814ffb44785', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8527f7733f44fd78cd259584ea40a99', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 763.367329] env[61907]: DEBUG oslo_vmware.api [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Task: {'id': task-1243646, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.113373} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.367329] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 763.367329] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 763.367329] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 763.367329] env[61907]: INFO nova.compute.manager [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Took 1.08 seconds to destroy the instance on the hypervisor. [ 763.367675] env[61907]: DEBUG oslo.service.loopingcall [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.367675] env[61907]: DEBUG nova.compute.manager [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 763.367675] env[61907]: DEBUG nova.network.neutron [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 763.382368] env[61907]: DEBUG nova.network.neutron [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.445060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.445060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.445060] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.676468] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 763.887025] env[61907]: DEBUG nova.network.neutron [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.930293] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Successfully created port: c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.000347] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.048348] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6adb034-a507-440c-9d7b-042c9d9d6156 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.057026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce02bb84-3978-4a16-9848-58596bbf5bbe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.088143] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4714657b-88e4-44f1-b667-45eb69c1bb8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.095497] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2d6af8a-3bff-4969-b769-bab11866a9d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.111162] env[61907]: DEBUG nova.compute.provider_tree [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.293717] env[61907]: DEBUG nova.network.neutron [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.389768] env[61907]: INFO nova.compute.manager [-] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Took 1.02 seconds to deallocate network for instance. [ 764.615172] env[61907]: DEBUG nova.scheduler.client.report [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 764.686496] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 764.712290] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 764.712896] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 764.713553] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 764.713797] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 764.713991] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 764.714195] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 764.714447] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 764.714680] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 764.714976] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 764.715836] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 764.716067] env[61907]: DEBUG nova.virt.hardware [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 764.716923] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddca8367-5a2f-4420-add1-affcb14ed0b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.727047] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d18d6b9b-fcf8-4b40-89be-7522ac45f7fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.796789] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.797154] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance network_info: |[{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 764.797675] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:94:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40a79714-91a5-4175-94fe-d7aeae0a8ea5', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 764.806558] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating folder: Project (3aab40034a4e4ccdb169abfc11f3b02f). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.806558] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-282cea8b-cc0e-4d6d-8565-44222cc51cee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.818247] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created folder: Project (3aab40034a4e4ccdb169abfc11f3b02f) in parent group-v268168. [ 764.818247] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating folder: Instances. Parent ref: group-v268188. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 764.819286] env[61907]: DEBUG nova.compute.manager [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 764.819462] env[61907]: DEBUG nova.compute.manager [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing instance network info cache due to event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 764.819673] env[61907]: DEBUG oslo_concurrency.lockutils [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.819809] env[61907]: DEBUG oslo_concurrency.lockutils [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.819958] env[61907]: DEBUG nova.network.neutron [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 764.821480] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f09d906a-c98d-46d7-8fa1-e05bd29d5910 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.830951] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created folder: Instances in parent group-v268188. [ 764.831203] env[61907]: DEBUG oslo.service.loopingcall [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.831391] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 764.831826] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-806c0dda-e786-4da6-b64f-97a834ae7782 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.851896] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 764.851896] env[61907]: value = "task-1243649" [ 764.851896] env[61907]: _type = "Task" [ 764.851896] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.859756] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243649, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.896114] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.120674] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.121223] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 765.124292] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.536s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.125600] env[61907]: INFO nova.compute.claims [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.361984] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243649, 'name': CreateVM_Task, 'duration_secs': 0.294576} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.366104] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 765.382714] env[61907]: DEBUG oslo_vmware.service [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e23025f2-8fb3-4304-a17a-1ac1bfd5f704 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.388781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.389000] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.389458] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 765.389721] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af5ebb3e-02e7-427c-a892-603c84a0bd81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.396378] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 765.396378] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f36dd9-5c9f-8785-9876-fa495629f1b5" [ 765.396378] env[61907]: _type = "Task" [ 765.396378] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.406942] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f36dd9-5c9f-8785-9876-fa495629f1b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.595188] env[61907]: DEBUG nova.network.neutron [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updated VIF entry in instance network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 765.595601] env[61907]: DEBUG nova.network.neutron [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.630179] env[61907]: DEBUG nova.compute.utils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.634124] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 765.634315] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 765.683982] env[61907]: DEBUG nova.policy [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da98c262c7fb41dea5785c9b89f7fdb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2048329d91224454805d73742252fc4d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.792217] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Successfully updated port: c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 765.908422] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.909046] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 765.909592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.910047] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.910657] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 765.910937] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e7063d6a-1cd8-4b4e-8c08-c80b3fa9fe9f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.922173] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 765.922173] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 765.922173] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5e94ba7-cfda-4e4c-94e7-3194a7225ede {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.928298] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d06a4398-ca80-4b4a-ad42-5c0913e69520 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.933888] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 765.933888] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a4eae8-6fd0-73e3-6931-a9f76059017c" [ 765.933888] env[61907]: _type = "Task" [ 765.933888] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.943491] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a4eae8-6fd0-73e3-6931-a9f76059017c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.999628] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Successfully created port: 8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 766.098428] env[61907]: DEBUG oslo_concurrency.lockutils [req-fe72cadc-1aa7-4e5f-88b0-1eb43dea483a req-a15fea3a-41a4-4a26-aa73-11ed07739667 service nova] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.135359] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 766.294768] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.294935] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquired lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.295099] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 766.449524] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Preparing fetch location {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 766.449866] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating directory with path [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.451135] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9de7015a-fd8c-45ee-84b1-8225f4b4edda {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.465398] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created directory with path [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.465604] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Fetch image to [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 766.466475] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloading image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk on the data store datastore1 {{(pid=61907) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 766.466691] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3fb6158-7aee-45e9-b852-6c4169e0fff2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.474117] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9701901c-4868-4c02-bec2-3c349cc9bb79 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.486863] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b914488-b216-4b00-bd5e-2397277cd0b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.522146] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4d21a5-37cf-40f4-af16-1228b45dc27d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.527847] env[61907]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-ff0f6ab7-ce9b-4a75-bafe-86b4ed0051e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.544093] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d816059-8f55-4234-80a9-68105990f148 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.551635] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a852a38-67d4-44af-ad18-17e1fa42972e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.582586] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fcf4c15-de99-41fe-8166-738569673eb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.585645] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloading image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to the data store datastore1 {{(pid=61907) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 766.592964] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c088afe8-1c8f-4f8f-98da-6218aff7f559 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.607941] env[61907]: DEBUG nova.compute.provider_tree [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.649480] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61907) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 766.867716] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.966969] env[61907]: DEBUG nova.compute.manager [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Received event network-vif-plugged-c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 766.967210] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Acquiring lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.967560] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.967715] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.967820] env[61907]: DEBUG nova.compute.manager [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] No waiting events found dispatching network-vif-plugged-c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 766.967974] env[61907]: WARNING nova.compute.manager [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Received unexpected event network-vif-plugged-c7b1cf36-0ee3-4d59-b61c-1af1799938f5 for instance with vm_state building and task_state spawning. [ 766.968145] env[61907]: DEBUG nova.compute.manager [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Received event network-changed-c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 766.968382] env[61907]: DEBUG nova.compute.manager [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Refreshing instance network info cache due to event network-changed-c7b1cf36-0ee3-4d59-b61c-1af1799938f5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 766.968595] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Acquiring lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.115376] env[61907]: DEBUG nova.scheduler.client.report [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 767.146789] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 767.185628] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 767.185865] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 767.186033] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 767.186225] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 767.186369] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 767.186532] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 767.186754] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 767.186910] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 767.187081] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 767.187245] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 767.187416] env[61907]: DEBUG nova.virt.hardware [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 767.188309] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a1595a-6f08-4ff7-9dee-8d2f6d2141b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.199382] env[61907]: DEBUG nova.network.neutron [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Updating instance_info_cache with network_info: [{"id": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "address": "fa:16:3e:ad:a9:12", "network": {"id": "24f065c5-cbdc-448d-8422-7d6d0ee82d62", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1331557243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8527f7733f44fd78cd259584ea40a99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b1cf36-0e", "ovs_interfaceid": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.203040] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e61d3f-7c01-479e-ac3d-fbe426a145af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.345549] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Completed reading data from the image iterator. {{(pid=61907) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 767.345549] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 767.492634] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloaded image file data fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk on the data store datastore1 {{(pid=61907) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 767.498358] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Caching image {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 767.498358] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copying Virtual Disk [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk to [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.498358] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac676ce2-0bbd-494e-8178-557ea70e9799 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.510021] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 767.510021] env[61907]: value = "task-1243650" [ 767.510021] env[61907]: _type = "Task" [ 767.510021] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.516677] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243650, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.624086] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.500s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.624713] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 767.630113] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.289s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.685376] env[61907]: DEBUG nova.compute.manager [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Received event network-vif-plugged-8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 767.685638] env[61907]: DEBUG oslo_concurrency.lockutils [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] Acquiring lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.685856] env[61907]: DEBUG oslo_concurrency.lockutils [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.686035] env[61907]: DEBUG oslo_concurrency.lockutils [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.686217] env[61907]: DEBUG nova.compute.manager [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] No waiting events found dispatching network-vif-plugged-8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 767.686382] env[61907]: WARNING nova.compute.manager [req-b0b5417b-68b5-4c52-9962-2e67be6fa0ad req-85819730-2e0a-4b48-9a0b-d6de8493abe0 service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Received unexpected event network-vif-plugged-8ed80238-7bda-433c-95f1-38554331b3de for instance with vm_state building and task_state spawning. [ 767.706417] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Releasing lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.706766] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Instance network_info: |[{"id": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "address": "fa:16:3e:ad:a9:12", "network": {"id": "24f065c5-cbdc-448d-8422-7d6d0ee82d62", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1331557243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8527f7733f44fd78cd259584ea40a99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b1cf36-0e", "ovs_interfaceid": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 767.707112] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Acquired lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.707294] env[61907]: DEBUG nova.network.neutron [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Refreshing network info cache for port c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 767.708643] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:a9:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4c8a5d7c-ee1f-4a41-94e4-db31e85a398d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c7b1cf36-0ee3-4d59-b61c-1af1799938f5', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 767.715973] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Creating folder: Project (b8527f7733f44fd78cd259584ea40a99). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.719181] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-acdd958e-fe15-45c8-9f4c-3039fa123c14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.721820] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Successfully updated port: 8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.735035] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Created folder: Project (b8527f7733f44fd78cd259584ea40a99) in parent group-v268168. [ 767.735035] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Creating folder: Instances. Parent ref: group-v268191. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 767.735035] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3d7100fb-74f3-4949-a428-a8f0946b5137 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.744433] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Created folder: Instances in parent group-v268191. [ 767.744433] env[61907]: DEBUG oslo.service.loopingcall [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 767.744433] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 767.744669] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a2597d18-591a-4fb1-91e7-033eb0e6b619 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.771796] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 767.771796] env[61907]: value = "task-1243653" [ 767.771796] env[61907]: _type = "Task" [ 767.771796] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.780143] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.993924] env[61907]: DEBUG nova.network.neutron [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Updated VIF entry in instance network info cache for port c7b1cf36-0ee3-4d59-b61c-1af1799938f5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 767.994320] env[61907]: DEBUG nova.network.neutron [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Updating instance_info_cache with network_info: [{"id": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "address": "fa:16:3e:ad:a9:12", "network": {"id": "24f065c5-cbdc-448d-8422-7d6d0ee82d62", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-1331557243-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b8527f7733f44fd78cd259584ea40a99", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4c8a5d7c-ee1f-4a41-94e4-db31e85a398d", "external-id": "cl2-zone-613", "segmentation_id": 613, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc7b1cf36-0e", "ovs_interfaceid": "c7b1cf36-0ee3-4d59-b61c-1af1799938f5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.020483] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243650, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.138621] env[61907]: DEBUG nova.compute.utils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.142030] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 768.142427] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 768.197107] env[61907]: DEBUG nova.policy [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84750bb751724a2f8e5ba3107f73defd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e6a77496c4f4c7ea060d39ad1504c53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.229663] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.229802] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.229949] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.287472] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.497591] env[61907]: DEBUG oslo_concurrency.lockutils [req-a79de194-8723-4cee-afc8-1923414089fb req-464f77fd-c5c4-4a41-8277-1eb1b6e27154 service nova] Releasing lock "refresh_cache-9209400d-d5b8-4b96-9022-f8a4c7a30471" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.518753] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96014831-5a3b-4032-88db-aa75d610f819 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.524894] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243650, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.68618} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.525692] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copied Virtual Disk [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk to [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 768.525944] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleting the datastore file [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/tmp-sparse.vmdk {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.526270] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-68507cb2-75d4-4a71-a0c4-d14084a15fe9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.530990] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78b9aa6-89f5-4b73-80c1-4a46c4d353af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.535530] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 768.535530] env[61907]: value = "task-1243654" [ 768.535530] env[61907]: _type = "Task" [ 768.535530] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.565654] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eafa423-beff-4352-bf90-6ab7fbba3d00 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.571145] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02265} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.571795] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 768.571942] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Moving file from [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 to [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57. {{(pid=61907) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 768.572208] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c2d4ea7c-11a5-470d-b7f7-e119b07eb6f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.577445] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb77a4d1-8409-487a-a566-82273ea7dced {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.582102] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 768.582102] env[61907]: value = "task-1243655" [ 768.582102] env[61907]: _type = "Task" [ 768.582102] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.599041] env[61907]: DEBUG nova.compute.provider_tree [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.599041] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243655, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.642675] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 768.696522] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Successfully created port: 05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.758551] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.785607] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.949541] env[61907]: DEBUG nova.network.neutron [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Updating instance_info_cache with network_info: [{"id": "8ed80238-7bda-433c-95f1-38554331b3de", "address": "fa:16:3e:3a:eb:c1", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ed80238-7b", "ovs_interfaceid": "8ed80238-7bda-433c-95f1-38554331b3de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.093241] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243655, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024042} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.093502] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] File moved {{(pid=61907) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 769.093697] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Cleaning up location [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 769.093874] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleting the datastore file [datastore1] vmware_temp/93410b17-36a6-4f20-95d2-a01424ad1c5f {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 769.094120] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddebb82b-66d0-4378-b67a-9b9270d21565 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.096702] env[61907]: DEBUG nova.scheduler.client.report [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 769.105723] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 769.105723] env[61907]: value = "task-1243656" [ 769.105723] env[61907]: _type = "Task" [ 769.105723] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.113570] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.289944] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.451873] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.452233] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Instance network_info: |[{"id": "8ed80238-7bda-433c-95f1-38554331b3de", "address": "fa:16:3e:3a:eb:c1", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ed80238-7b", "ovs_interfaceid": "8ed80238-7bda-433c-95f1-38554331b3de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 769.452652] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:eb:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '744515ee-aa5b-4c23-b959-b56c51da6b86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ed80238-7bda-433c-95f1-38554331b3de', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.459953] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Creating folder: Project (2048329d91224454805d73742252fc4d). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.460591] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a173dd53-cccf-4de6-901e-d1dec9aa0127 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.470823] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Created folder: Project (2048329d91224454805d73742252fc4d) in parent group-v268168. [ 769.471048] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Creating folder: Instances. Parent ref: group-v268194. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 769.471731] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6c5f284c-cda4-4e9f-b8e6-10fee851d75e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.480247] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Created folder: Instances in parent group-v268194. [ 769.480560] env[61907]: DEBUG oslo.service.loopingcall [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.480707] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 769.481066] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c2f69c1f-a951-4a68-9bdd-1477940da982 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.501271] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.501271] env[61907]: value = "task-1243659" [ 769.501271] env[61907]: _type = "Task" [ 769.501271] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.511336] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243659, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.601918] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.603177] env[61907]: ERROR nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Traceback (most recent call last): [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.driver.spawn(context, instance, image_meta, [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] vm_ref = self.build_virtual_machine(instance, [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] vif_infos = vmwarevif.get_vif_info(self._session, [ 769.603177] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] for vif in network_info: [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return self._sync_wrapper(fn, *args, **kwargs) [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.wait() [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self[:] = self._gt.wait() [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return self._exit_event.wait() [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] current.throw(*self._exc) [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 769.603513] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] result = function(*args, **kwargs) [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] return func(*args, **kwargs) [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise e [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] nwinfo = self.network_api.allocate_for_instance( [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] created_port_ids = self._update_ports_for_instance( [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] with excutils.save_and_reraise_exception(): [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] self.force_reraise() [ 769.603866] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise self.value [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] updated_port = self._update_port( [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] _ensure_no_port_binding_failure(port) [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] raise exception.PortBindingFailed(port_id=port['id']) [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] nova.exception.PortBindingFailed: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. [ 769.604254] env[61907]: ERROR nova.compute.manager [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] [ 769.604254] env[61907]: DEBUG nova.compute.utils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.605663] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.345s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.605663] env[61907]: DEBUG nova.objects.instance [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 769.607353] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Build of instance 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d was re-scheduled: Binding failed for port c9747636-55a8-49ff-ac51-d00435e9ba44, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 769.607783] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 769.608018] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquiring lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.608169] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Acquired lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.608329] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.620047] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024451} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.620301] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.621107] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfc5cf06-37d0-412f-97ee-27497f429e3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.626357] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 769.626357] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cbbc5d-b121-f22a-a859-9446467a1b45" [ 769.626357] env[61907]: _type = "Task" [ 769.626357] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.638584] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cbbc5d-b121-f22a-a859-9446467a1b45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.652089] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 769.682934] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 769.683316] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 769.683561] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 769.683810] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 769.684665] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 769.684665] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 769.684665] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 769.684665] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 769.684665] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 769.685042] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 769.685222] env[61907]: DEBUG nova.virt.hardware [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.686127] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a2b2397-93d2-4598-be7f-f5aca4497c1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.694613] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f24f65-9428-437b-b4f9-f7af85ea1427 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.715422] env[61907]: DEBUG nova.compute.manager [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Received event network-changed-8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 769.715653] env[61907]: DEBUG nova.compute.manager [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Refreshing instance network info cache due to event network-changed-8ed80238-7bda-433c-95f1-38554331b3de. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 769.715879] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] Acquiring lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.716345] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] Acquired lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.716596] env[61907]: DEBUG nova.network.neutron [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Refreshing network info cache for port 8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 769.788480] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.010951] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243659, 'name': CreateVM_Task, 'duration_secs': 0.342643} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.011146] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 770.011924] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.012132] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.012394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 770.012680] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c287ac3b-f293-478a-8baa-c02394df3d75 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.017294] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 770.017294] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527857a6-a98a-1412-34ca-7815ba547f32" [ 770.017294] env[61907]: _type = "Task" [ 770.017294] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.024869] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527857a6-a98a-1412-34ca-7815ba547f32, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.131839] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.138851] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cbbc5d-b121-f22a-a859-9446467a1b45, 'name': SearchDatastore_Task, 'duration_secs': 0.013851} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.140080] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.140337] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 770.142907] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7481552-70b4-4707-ad93-29b6a60d7983 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.145467] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.145682] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.150187] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 770.150187] env[61907]: value = "task-1243660" [ 770.150187] env[61907]: _type = "Task" [ 770.150187] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.157749] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243660, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.207165] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.287616] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.519988] env[61907]: DEBUG nova.network.neutron [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Updated VIF entry in instance network info cache for port 8ed80238-7bda-433c-95f1-38554331b3de. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 770.520372] env[61907]: DEBUG nova.network.neutron [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Updating instance_info_cache with network_info: [{"id": "8ed80238-7bda-433c-95f1-38554331b3de", "address": "fa:16:3e:3a:eb:c1", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ed80238-7b", "ovs_interfaceid": "8ed80238-7bda-433c-95f1-38554331b3de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.532214] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527857a6-a98a-1412-34ca-7815ba547f32, 'name': SearchDatastore_Task, 'duration_secs': 0.011185} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.532515] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.532756] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.533160] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.533341] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.533630] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.534644] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c537aa60-e1d8-4b28-b3e6-1dbfecdbfb09 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.539600] env[61907]: DEBUG nova.compute.manager [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Received event network-vif-plugged-05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 770.539795] env[61907]: DEBUG oslo_concurrency.lockutils [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] Acquiring lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.540111] env[61907]: DEBUG oslo_concurrency.lockutils [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.540351] env[61907]: DEBUG oslo_concurrency.lockutils [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.540623] env[61907]: DEBUG nova.compute.manager [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] No waiting events found dispatching network-vif-plugged-05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 770.540806] env[61907]: WARNING nova.compute.manager [req-c0228074-5c14-4560-a005-bff853b2dab9 req-a39a5f00-11fd-4c22-a3b0-3641ba58795a service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Received unexpected event network-vif-plugged-05a00a60-c277-47b8-98e8-bc8f0f46258a for instance with vm_state building and task_state spawning. [ 770.555032] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.555246] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 770.556452] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-896664c1-c55a-4263-9504-0336ce765b22 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.563566] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Successfully updated port: 05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 770.570346] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 770.570346] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525c24bb-f680-6934-2b01-e6caacb16960" [ 770.570346] env[61907]: _type = "Task" [ 770.570346] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.585137] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525c24bb-f680-6934-2b01-e6caacb16960, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.617212] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bd87f0e3-4569-453c-bceb-736c8a8a1dc3 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.619839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.656s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.621072] env[61907]: INFO nova.compute.claims [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 770.660933] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243660, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477056} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.661374] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 770.661528] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.661786] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0fb717c1-d5e8-4f80-b3f6-e46fe86d8509 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.669433] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 770.669433] env[61907]: value = "task-1243661" [ 770.669433] env[61907]: _type = "Task" [ 770.669433] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.677225] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243661, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.710149] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Releasing lock "refresh_cache-1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.710489] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 770.710794] env[61907]: DEBUG nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 770.711045] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.731629] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.785813] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.026222] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad38323d-19c5-4c0a-b626-937b49e81d3c req-381f9291-57e4-4445-83bf-5085d80925ae service nova] Releasing lock "refresh_cache-a6a117b3-7cc4-43c5-8202-12514f695e48" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.066264] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.066709] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.066709] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 771.081538] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525c24bb-f680-6934-2b01-e6caacb16960, 'name': SearchDatastore_Task, 'duration_secs': 0.043267} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.082820] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c46fb994-946e-4f1f-a86b-a993305f4866 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.087445] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 771.087445] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ec777e-b307-0793-bbbe-908374418f23" [ 771.087445] env[61907]: _type = "Task" [ 771.087445] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.095389] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ec777e-b307-0793-bbbe-908374418f23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.180580] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243661, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061761} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.180865] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 771.181655] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867473de-717a-483e-a406-54e727c189e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.203134] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.203651] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-956afa4b-554a-4334-aaae-22fe0a4781af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.222489] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 771.222489] env[61907]: value = "task-1243662" [ 771.222489] env[61907]: _type = "Task" [ 771.222489] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.229872] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243662, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.234348] env[61907]: DEBUG nova.network.neutron [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.287712] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.598291] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ec777e-b307-0793-bbbe-908374418f23, 'name': SearchDatastore_Task, 'duration_secs': 0.009124} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.599154] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.601104] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.601384] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] a6a117b3-7cc4-43c5-8202-12514f695e48/a6a117b3-7cc4-43c5-8202-12514f695e48.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 771.601670] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a7c895f6-13b6-48a0-96a1-f9f9489c0ec4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.608606] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 771.608606] env[61907]: value = "task-1243663" [ 771.608606] env[61907]: _type = "Task" [ 771.608606] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.616753] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243663, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.738895] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243662, 'name': ReconfigVM_Task, 'duration_secs': 0.326398} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.738895] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.738895] env[61907]: INFO nova.compute.manager [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] [instance: 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d] Took 1.03 seconds to deallocate network for instance. [ 771.740103] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-660e1d2e-4408-4838-a57d-e203390ca759 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.748862] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 771.748862] env[61907]: value = "task-1243664" [ 771.748862] env[61907]: _type = "Task" [ 771.748862] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.758499] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243664, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.765156] env[61907]: DEBUG nova.network.neutron [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [{"id": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "address": "fa:16:3e:0f:68:53", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05a00a60-c2", "ovs_interfaceid": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.793625] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243653, 'name': CreateVM_Task, 'duration_secs': 3.581604} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.796428] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 771.798240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.798240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.798240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 771.798600] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9587e31f-f9c7-45eb-88ad-a25937686829 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.805611] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 771.805611] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292acdc-fae8-539b-df26-1561dc94fc8e" [ 771.805611] env[61907]: _type = "Task" [ 771.805611] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.818783] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292acdc-fae8-539b-df26-1561dc94fc8e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.979515] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a7b11f-6cc9-4222-9d29-ff1732c585dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.988674] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2feaa875-fb76-4a9a-911e-ecee915017f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.022442] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea30b61-8b8d-490f-8312-62f41b1b4986 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.030307] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-063e43ef-ad3b-4b87-88a6-74ad3f72f17d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.045758] env[61907]: DEBUG nova.compute.provider_tree [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.119145] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243663, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.119371] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] a6a117b3-7cc4-43c5-8202-12514f695e48/a6a117b3-7cc4-43c5-8202-12514f695e48.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 772.119589] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.119879] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f1e5ea80-8ec0-4607-bc6d-c6f5574b018c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.125673] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 772.125673] env[61907]: value = "task-1243665" [ 772.125673] env[61907]: _type = "Task" [ 772.125673] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.133351] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243665, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.257736] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243664, 'name': Rename_Task, 'duration_secs': 0.16716} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.258014] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 772.258256] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cce726d2-740c-4260-b27e-8f1d22e210a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.263831] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 772.263831] env[61907]: value = "task-1243666" [ 772.263831] env[61907]: _type = "Task" [ 772.263831] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.273036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.273036] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Instance network_info: |[{"id": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "address": "fa:16:3e:0f:68:53", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05a00a60-c2", "ovs_interfaceid": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 772.273253] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243666, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.273253] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0f:68:53', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '05a00a60-c277-47b8-98e8-bc8f0f46258a', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.279774] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating folder: Project (8e6a77496c4f4c7ea060d39ad1504c53). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.280699] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-244d7858-bdb4-4efb-90af-e87c5af5fe19 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.289594] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created folder: Project (8e6a77496c4f4c7ea060d39ad1504c53) in parent group-v268168. [ 772.289779] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating folder: Instances. Parent ref: group-v268197. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 772.290009] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2d7583a-5528-4796-9beb-61774fba4eaf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.299584] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created folder: Instances in parent group-v268197. [ 772.299824] env[61907]: DEBUG oslo.service.loopingcall [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.300026] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.300271] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8e6c4c6-101d-44e5-b04a-ceba51d08767 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.323185] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292acdc-fae8-539b-df26-1561dc94fc8e, 'name': SearchDatastore_Task, 'duration_secs': 0.049989} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.324450] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.324692] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.325114] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.325235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.325410] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.325687] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.325687] env[61907]: value = "task-1243669" [ 772.325687] env[61907]: _type = "Task" [ 772.325687] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.325940] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-42b5b18e-ff1c-4a4b-8b28-d9de92fadef8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.335292] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243669, 'name': CreateVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.336335] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.336508] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.337225] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33ad887d-4804-4337-a210-46e897055604 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.341640] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 772.341640] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d21505-d7d4-27b7-6c2c-d87d325e4ce8" [ 772.341640] env[61907]: _type = "Task" [ 772.341640] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.350474] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d21505-d7d4-27b7-6c2c-d87d325e4ce8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.551017] env[61907]: DEBUG nova.scheduler.client.report [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 772.570201] env[61907]: DEBUG nova.compute.manager [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Received event network-changed-05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 772.570437] env[61907]: DEBUG nova.compute.manager [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Refreshing instance network info cache due to event network-changed-05a00a60-c277-47b8-98e8-bc8f0f46258a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 772.570708] env[61907]: DEBUG oslo_concurrency.lockutils [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] Acquiring lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.570894] env[61907]: DEBUG oslo_concurrency.lockutils [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] Acquired lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.571130] env[61907]: DEBUG nova.network.neutron [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Refreshing network info cache for port 05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.636160] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243665, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092329} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.636482] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.637217] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99eec2d9-ff28-41f4-bef5-7f35a39c1d61 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.660897] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] a6a117b3-7cc4-43c5-8202-12514f695e48/a6a117b3-7cc4-43c5-8202-12514f695e48.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.661155] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28bd4dbe-b92f-41be-95a5-58e611644639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.680679] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 772.680679] env[61907]: value = "task-1243670" [ 772.680679] env[61907]: _type = "Task" [ 772.680679] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.688758] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243670, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.770393] env[61907]: INFO nova.scheduler.client.report [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Deleted allocations for instance 1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d [ 772.779659] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.840864] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243669, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.851794] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d21505-d7d4-27b7-6c2c-d87d325e4ce8, 'name': SearchDatastore_Task, 'duration_secs': 0.008112} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.852684] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-effe828b-5d6f-4207-9668-b97528d78d5e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.857861] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 772.857861] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e975d-ac4d-7d33-0f06-24ff89799de5" [ 772.857861] env[61907]: _type = "Task" [ 772.857861] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.866294] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e975d-ac4d-7d33-0f06-24ff89799de5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.055218] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.055783] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 773.058786] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.286s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.060148] env[61907]: INFO nova.compute.claims [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.192329] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243670, 'name': ReconfigVM_Task, 'duration_secs': 0.315591} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.192598] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Reconfigured VM instance instance-00000032 to attach disk [datastore1] a6a117b3-7cc4-43c5-8202-12514f695e48/a6a117b3-7cc4-43c5-8202-12514f695e48.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.193217] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-404647c5-f4e3-41bf-b917-2b042ffab205 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.199143] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 773.199143] env[61907]: value = "task-1243671" [ 773.199143] env[61907]: _type = "Task" [ 773.199143] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.209333] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243671, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.258613] env[61907]: DEBUG nova.network.neutron [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updated VIF entry in instance network info cache for port 05a00a60-c277-47b8-98e8-bc8f0f46258a. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 773.258613] env[61907]: DEBUG nova.network.neutron [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [{"id": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "address": "fa:16:3e:0f:68:53", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05a00a60-c2", "ovs_interfaceid": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.276101] env[61907]: DEBUG oslo_vmware.api [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243666, 'name': PowerOnVM_Task, 'duration_secs': 0.865841} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.276376] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 773.276788] env[61907]: INFO nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Took 11.50 seconds to spawn the instance on the hypervisor. [ 773.277065] env[61907]: DEBUG nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 773.277963] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69679433-dd0c-45c9-9bf6-1e0df7296ef8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.284640] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d0da2e8c-6c2b-41dc-8012-bc4e59e0dfa3 tempest-ServersNegativeTestJSON-892224878 tempest-ServersNegativeTestJSON-892224878-project-member] Lock "1f8c2103-0a62-4be6-b4e6-3d5f3a787f4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.792s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.343122] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243669, 'name': CreateVM_Task, 'duration_secs': 0.620684} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.343293] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.343993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.344167] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.344479] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.344734] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f41ccbaa-631a-4b79-bc98-bd610a582332 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.349214] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 773.349214] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521120ae-f87f-eec0-8fb1-19285b0f3442" [ 773.349214] env[61907]: _type = "Task" [ 773.349214] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.357085] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521120ae-f87f-eec0-8fb1-19285b0f3442, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.366201] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e975d-ac4d-7d33-0f06-24ff89799de5, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.366449] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.366703] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 9209400d-d5b8-4b96-9022-f8a4c7a30471/9209400d-d5b8-4b96-9022-f8a4c7a30471.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.366966] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c109fd2c-2ac7-4598-b6e0-4fe09179163e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.374122] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 773.374122] env[61907]: value = "task-1243672" [ 773.374122] env[61907]: _type = "Task" [ 773.374122] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.381727] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.560593] env[61907]: DEBUG nova.compute.utils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 773.562384] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 773.562384] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 773.612990] env[61907]: DEBUG nova.policy [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.710244] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243671, 'name': Rename_Task, 'duration_secs': 0.137947} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.710667] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 773.712027] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6dea252-21e7-4ca7-9673-ad4a765a9245 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.718390] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 773.718390] env[61907]: value = "task-1243673" [ 773.718390] env[61907]: _type = "Task" [ 773.718390] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.732035] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243673, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.761225] env[61907]: DEBUG oslo_concurrency.lockutils [req-de4c08d5-77ff-42b4-9861-abde96524053 req-062b669f-54ed-4f6a-b668-b9d30e850f8c service nova] Releasing lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.788273] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 773.809697] env[61907]: INFO nova.compute.manager [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Took 29.86 seconds to build instance. [ 773.864065] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521120ae-f87f-eec0-8fb1-19285b0f3442, 'name': SearchDatastore_Task, 'duration_secs': 0.008096} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.864065] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.864065] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.864065] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.864699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.864699] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.864699] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43aa08cc-fc7d-4862-92a2-f7f6cb930ac4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.873753] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.873939] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 773.877805] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cad0e4e-eabf-47df-9a0a-d2e9681d340a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.884399] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 773.884399] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fb8180-a694-4cac-ba4b-863598a91b35" [ 773.884399] env[61907]: _type = "Task" [ 773.884399] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.888115] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243672, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500666} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.891986] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 9209400d-d5b8-4b96-9022-f8a4c7a30471/9209400d-d5b8-4b96-9022-f8a4c7a30471.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 773.892277] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 773.892776] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e988cbc-c8f4-4db7-9c18-0eb9d90b8951 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.901500] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fb8180-a694-4cac-ba4b-863598a91b35, 'name': SearchDatastore_Task, 'duration_secs': 0.012622} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.903811] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 773.903811] env[61907]: value = "task-1243674" [ 773.903811] env[61907]: _type = "Task" [ 773.903811] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.904286] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8c6fd03-05c5-4ca2-8523-0c4569c8acff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.912507] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 773.912507] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526510f4-87c0-4266-acb3-89beecdb7bd1" [ 773.912507] env[61907]: _type = "Task" [ 773.912507] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.915640] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243674, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.924211] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526510f4-87c0-4266-acb3-89beecdb7bd1, 'name': SearchDatastore_Task, 'duration_secs': 0.008088} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.924498] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.924783] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/865d386d-c5f0-439b-b50b-b613993ec9b8.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.925089] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e2d5c793-c934-4c57-8773-696beefbdc58 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.931439] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 773.931439] env[61907]: value = "task-1243675" [ 773.931439] env[61907]: _type = "Task" [ 773.931439] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.941027] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243675, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.964897] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Successfully created port: 1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.070028] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 774.231646] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243673, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.312516] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3936606-e18d-4d40-9a0c-2716e7036272 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.050s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.326791] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.422199] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243674, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069063} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.422522] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 774.423400] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df71a3e8-99b1-40ec-b0c6-1ee6eabedf4e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.455760] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 9209400d-d5b8-4b96-9022-f8a4c7a30471/9209400d-d5b8-4b96-9022-f8a4c7a30471.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 774.462247] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3800395-f232-4cb0-9489-41344a4fc9ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.482885] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243675, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.484253] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 774.484253] env[61907]: value = "task-1243676" [ 774.484253] env[61907]: _type = "Task" [ 774.484253] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.494470] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243676, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.502677] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc595f2-30a4-4dad-a1ad-2c4eb34c238e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.509716] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aadaff7-919e-4616-9602-e519376f9bc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.542018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11edbc7d-2a07-466b-ad88-b2d8060c8dc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.549091] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3e29f1-96d9-49ec-9ea9-645ba9ad2fa0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.562146] env[61907]: DEBUG nova.compute.provider_tree [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.729480] env[61907]: DEBUG oslo_vmware.api [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243673, 'name': PowerOnVM_Task, 'duration_secs': 0.553842} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.729740] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 774.729940] env[61907]: INFO nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Took 7.58 seconds to spawn the instance on the hypervisor. [ 774.730131] env[61907]: DEBUG nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 774.730996] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4cc5c8-5bbc-439e-8617-5b174d9ff941 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.815363] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 774.955561] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243675, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582482} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.955823] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/865d386d-c5f0-439b-b50b-b613993ec9b8.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.956049] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.956299] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-da5c86b1-5ad4-41cb-a521-96310d9aba5c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.963947] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 774.963947] env[61907]: value = "task-1243677" [ 774.963947] env[61907]: _type = "Task" [ 774.963947] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.972686] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243677, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.993750] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.064979] env[61907]: DEBUG nova.scheduler.client.report [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 775.084288] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 775.110105] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.110374] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.110524] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.110842] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.110915] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.111063] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.111240] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.111399] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.111561] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.111729] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.111941] env[61907]: DEBUG nova.virt.hardware [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.112748] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8803de6-5b81-42f5-917b-1aaab4c09ede {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.121113] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489f0deb-0cd9-4de4-8eef-46321e2947ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.254337] env[61907]: INFO nova.compute.manager [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Took 29.57 seconds to build instance. [ 775.345110] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.393800] env[61907]: DEBUG nova.compute.manager [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 775.393934] env[61907]: DEBUG nova.compute.manager [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing instance network info cache due to event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 775.394200] env[61907]: DEBUG oslo_concurrency.lockutils [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.394377] env[61907]: DEBUG oslo_concurrency.lockutils [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.394780] env[61907]: DEBUG nova.network.neutron [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 775.473557] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243677, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063486} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.473879] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.474827] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd128613-e840-4a04-9183-43a76ecd989d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.502783] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/865d386d-c5f0-439b-b50b-b613993ec9b8.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.507195] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd5aa42a-c9da-4274-a497-0f60ccfdc4bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.528865] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243676, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.530340] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 775.530340] env[61907]: value = "task-1243678" [ 775.530340] env[61907]: _type = "Task" [ 775.530340] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.539660] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243678, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.570093] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.570732] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 775.573468] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.636s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.724428] env[61907]: DEBUG nova.compute.manager [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received event network-vif-plugged-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 775.724645] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.724887] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.725243] env[61907]: DEBUG oslo_concurrency.lockutils [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.725438] env[61907]: DEBUG nova.compute.manager [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] No waiting events found dispatching network-vif-plugged-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 775.725658] env[61907]: WARNING nova.compute.manager [req-bf5c529d-7872-4325-831c-e36811845297 req-fefdcfd5-a6c1-4359-a26d-b17717b5b134 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received unexpected event network-vif-plugged-1994ca95-0956-444e-8bb0-cd6a773e7ede for instance with vm_state building and task_state spawning. [ 775.758904] env[61907]: DEBUG oslo_concurrency.lockutils [None req-715736fb-a4a6-479e-8786-1aad3988dd72 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.210s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.906394] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Successfully updated port: 1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.008667] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243676, 'name': ReconfigVM_Task, 'duration_secs': 1.28877} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.008667] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 9209400d-d5b8-4b96-9022-f8a4c7a30471/9209400d-d5b8-4b96-9022-f8a4c7a30471.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.009380] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5044bec4-030d-46a2-ad5f-00bcb7823316 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.018450] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 776.018450] env[61907]: value = "task-1243679" [ 776.018450] env[61907]: _type = "Task" [ 776.018450] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.028475] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243679, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.040864] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243678, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.081034] env[61907]: DEBUG nova.compute.utils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.089115] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 776.089439] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 776.207081] env[61907]: DEBUG nova.policy [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41fc76e7ed9a46a09f96f11b28556ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b57c1194e0f14ae498d99d302da346c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.240122] env[61907]: DEBUG nova.network.neutron [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updated VIF entry in instance network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 776.240122] env[61907]: DEBUG nova.network.neutron [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.262042] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 776.410405] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.410405] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.410405] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 776.532174] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243679, 'name': Rename_Task, 'duration_secs': 0.197895} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.537874] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.538144] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbefc6c7-2bd5-4fb3-85b8-f142299c50e1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.544773] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243678, 'name': ReconfigVM_Task, 'duration_secs': 0.704842} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.546119] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/865d386d-c5f0-439b-b50b-b613993ec9b8.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.546759] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 776.546759] env[61907]: value = "task-1243680" [ 776.546759] env[61907]: _type = "Task" [ 776.546759] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.547078] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60ccfe42-e24b-4260-ac91-ea3bb212e4d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.556926] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243680, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.558172] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 776.558172] env[61907]: value = "task-1243681" [ 776.558172] env[61907]: _type = "Task" [ 776.558172] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.568451] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243681, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.590014] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 776.623658] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance da7ddc92-f3e3-4b4d-a14b-13fd25a72314 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 776.623848] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 215d0f35-5d19-4db4-996d-571e2970e19c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.623979] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 45c58b31-d0cc-4b36-b744-448a38e11266 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624119] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624249] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 9209400d-d5b8-4b96-9022-f8a4c7a30471 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624366] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a6a117b3-7cc4-43c5-8202-12514f695e48 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624492] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 865d386d-c5f0-439b-b50b-b613993ec9b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624603] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 3c9ee76a-464d-437e-9b73-862f1da24f7e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.624713] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 1ef8594c-5d9a-491f-89f6-e5d7f35ce751 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 776.743500] env[61907]: DEBUG oslo_concurrency.lockutils [req-8a8c23ad-b1f2-4f2f-ab60-6c8f098cf499 req-75373840-458f-4abb-93fb-934a0041635d service nova] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.789302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.912436] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Successfully created port: 72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 776.962667] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 777.062982] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243680, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.072287] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243681, 'name': Rename_Task, 'duration_secs': 0.176892} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.072580] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.072814] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17bdc953-2581-4e81-ba37-8367fa58e62e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.080110] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 777.080110] env[61907]: value = "task-1243682" [ 777.080110] env[61907]: _type = "Task" [ 777.080110] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.088895] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243682, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.129215] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5b20f363-dfad-4c2b-b757-4da4e5bd0b99 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.171121] env[61907]: DEBUG nova.network.neutron [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updating instance_info_cache with network_info: [{"id": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "address": "fa:16:3e:d1:48:17", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1994ca95-09", "ovs_interfaceid": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.561311] env[61907]: DEBUG oslo_vmware.api [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243680, 'name': PowerOnVM_Task, 'duration_secs': 0.595154} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.561565] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.561811] env[61907]: INFO nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Took 12.88 seconds to spawn the instance on the hypervisor. [ 777.562016] env[61907]: DEBUG nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 777.562785] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6065ecbc-f845-4841-ac2a-9889c9835c52 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.591185] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243682, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.599575] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 777.625685] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 777.625941] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 777.626176] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 777.626390] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 777.626535] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 777.626682] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 777.626883] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 777.627059] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 777.627232] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 777.627392] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 777.627565] env[61907]: DEBUG nova.virt.hardware [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 777.628537] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227cbd84-da22-44e4-92a1-03f6a6c7e99b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.632676] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 08647e48-9e24-4d7c-a9fd-4066317a7250 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 777.637079] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d361671-064b-4d2b-9627-781d894770b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.674943] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.675280] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance network_info: |[{"id": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "address": "fa:16:3e:d1:48:17", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1994ca95-09", "ovs_interfaceid": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 777.675718] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:48:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1994ca95-0956-444e-8bb0-cd6a773e7ede', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 777.683118] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating folder: Project (d8b341077bf045e0b079bbe8aa08cae0). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.683391] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ab98b75b-6dd1-4b2c-8ac2-6d746049fb17 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.694188] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created folder: Project (d8b341077bf045e0b079bbe8aa08cae0) in parent group-v268168. [ 777.694439] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating folder: Instances. Parent ref: group-v268200. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 777.694723] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d3082b87-cd7c-4eb8-ae2a-b9262ed284ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.703772] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created folder: Instances in parent group-v268200. [ 777.704092] env[61907]: DEBUG oslo.service.loopingcall [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.704380] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 777.704610] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8a7fdb3-8d6e-495a-bca9-2f90b2100ef4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.723193] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 777.723193] env[61907]: value = "task-1243685" [ 777.723193] env[61907]: _type = "Task" [ 777.723193] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.730887] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243685, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.753624] env[61907]: DEBUG nova.compute.manager [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received event network-changed-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 777.753624] env[61907]: DEBUG nova.compute.manager [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Refreshing instance network info cache due to event network-changed-1994ca95-0956-444e-8bb0-cd6a773e7ede. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 777.753624] env[61907]: DEBUG oslo_concurrency.lockutils [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] Acquiring lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.753624] env[61907]: DEBUG oslo_concurrency.lockutils [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] Acquired lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.753624] env[61907]: DEBUG nova.network.neutron [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Refreshing network info cache for port 1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 778.081980] env[61907]: INFO nova.compute.manager [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Took 33.97 seconds to build instance. [ 778.092748] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243682, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.137921] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ce864af6-aef6-4044-be64-8440ba175438 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 778.232358] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243685, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.550778] env[61907]: DEBUG nova.network.neutron [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updated VIF entry in instance network info cache for port 1994ca95-0956-444e-8bb0-cd6a773e7ede. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 778.551154] env[61907]: DEBUG nova.network.neutron [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updating instance_info_cache with network_info: [{"id": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "address": "fa:16:3e:d1:48:17", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1994ca95-09", "ovs_interfaceid": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.584275] env[61907]: DEBUG oslo_concurrency.lockutils [None req-45742a7d-d5f6-4fa3-a2b8-2100eeb84f52 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.351s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.594391] env[61907]: DEBUG oslo_vmware.api [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243682, 'name': PowerOnVM_Task, 'duration_secs': 1.302446} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.594717] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.594974] env[61907]: INFO nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Took 8.94 seconds to spawn the instance on the hypervisor. [ 778.595818] env[61907]: DEBUG nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 778.595974] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e2e26c6-7cd8-4fa8-8673-9429359e6a8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.642080] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Successfully updated port: 72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 778.646592] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 52404147-0375-4f9c-9e1e-4a52001df7cb has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 778.733607] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243685, 'name': CreateVM_Task, 'duration_secs': 0.687159} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.733785] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 778.734503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.734677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.735156] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 778.735410] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3573be04-cf9f-4ba1-853e-884633a939e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.739994] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 778.739994] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292d6ee-a824-118e-880b-f0a0dececd1c" [ 778.739994] env[61907]: _type = "Task" [ 778.739994] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.748086] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292d6ee-a824-118e-880b-f0a0dececd1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.054746] env[61907]: DEBUG oslo_concurrency.lockutils [req-6d3c636f-6cfd-4c92-bd6b-002aaa5bbd14 req-a369c9b3-905e-4110-bd47-484f3cd0f601 service nova] Releasing lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.090250] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 779.116677] env[61907]: INFO nova.compute.manager [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Took 31.56 seconds to build instance. [ 779.152308] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.152308] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.152308] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.152308] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f157c0db-054b-4d99-803c-11aa40cf0670 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 779.250594] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5292d6ee-a824-118e-880b-f0a0dececd1c, 'name': SearchDatastore_Task, 'duration_secs': 0.014287} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.250772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.251091] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.251299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.251400] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.251573] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.251837] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d764ce26-95f5-4879-9c03-b8472c619816 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.263752] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.264021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 779.265467] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d9bc4eb-10be-4d14-96b4-3ebcddbb30ec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.272383] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 779.272383] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ef8a8b-2f77-3759-1195-b3405da3dd49" [ 779.272383] env[61907]: _type = "Task" [ 779.272383] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.277633] env[61907]: DEBUG nova.compute.manager [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 779.278574] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ed99814-a8ba-455a-884e-a8564442e33c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.285382] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ef8a8b-2f77-3759-1195-b3405da3dd49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.615839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.619279] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5a1a98fb-31ff-481c-8c9a-aba856c7c82f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.283s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.661495] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 779.703516] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 779.758650] env[61907]: INFO nova.compute.manager [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Rescuing [ 779.759048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.759244] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.759438] env[61907]: DEBUG nova.network.neutron [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.783798] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ef8a8b-2f77-3759-1195-b3405da3dd49, 'name': SearchDatastore_Task, 'duration_secs': 0.009071} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.784637] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e446cacd-bf3b-438f-862d-b17c37b33e31 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.789989] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 779.789989] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521ae6e0-5e81-0e42-9e46-ae0368ebf04e" [ 779.789989] env[61907]: _type = "Task" [ 779.789989] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.794092] env[61907]: INFO nova.compute.manager [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] instance snapshotting [ 779.797982] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d0d893-b08c-4bf2-8800-029b3cad3f33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.804745] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521ae6e0-5e81-0e42-9e46-ae0368ebf04e, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.805377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.805660] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 3c9ee76a-464d-437e-9b73-862f1da24f7e/3c9ee76a-464d-437e-9b73-862f1da24f7e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 779.805936] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8db6ab51-d830-4057-a8a1-69448e324671 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.825216] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27e3369-8f9b-4492-a0e1-998e706548d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.831143] env[61907]: DEBUG nova.compute.manager [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Received event network-vif-plugged-72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 779.831311] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Acquiring lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.831460] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.831642] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.831870] env[61907]: DEBUG nova.compute.manager [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] No waiting events found dispatching network-vif-plugged-72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 779.832030] env[61907]: WARNING nova.compute.manager [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Received unexpected event network-vif-plugged-72964168-fdd5-43bb-9940-86f4349e333b for instance with vm_state building and task_state spawning. [ 779.832211] env[61907]: DEBUG nova.compute.manager [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Received event network-changed-72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 779.832376] env[61907]: DEBUG nova.compute.manager [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Refreshing instance network info cache due to event network-changed-72964168-fdd5-43bb-9940-86f4349e333b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 779.832553] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Acquiring lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.833163] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 779.833163] env[61907]: value = "task-1243686" [ 779.833163] env[61907]: _type = "Task" [ 779.833163] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.846514] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243686, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.913902] env[61907]: DEBUG nova.network.neutron [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Updating instance_info_cache with network_info: [{"id": "72964168-fdd5-43bb-9940-86f4349e333b", "address": "fa:16:3e:a5:4e:a7", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72964168-fd", "ovs_interfaceid": "72964168-fdd5-43bb-9940-86f4349e333b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.122023] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 780.166158] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e5d4890e-be1b-4b28-9721-89ef678c1182 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 780.343140] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 780.343456] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243686, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435521} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.345695] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-82501575-30f5-41a1-b03a-9260d0e7cdf3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.349197] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 3c9ee76a-464d-437e-9b73-862f1da24f7e/3c9ee76a-464d-437e-9b73-862f1da24f7e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 780.349405] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 780.349627] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4ed4742-bf17-40ad-ad88-122c145afeb9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.355999] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 780.355999] env[61907]: value = "task-1243688" [ 780.355999] env[61907]: _type = "Task" [ 780.355999] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.357285] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 780.357285] env[61907]: value = "task-1243687" [ 780.357285] env[61907]: _type = "Task" [ 780.357285] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.133308] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.133616] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.133759] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.133933] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.134112] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.135763] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.136066] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance network_info: |[{"id": "72964168-fdd5-43bb-9940-86f4349e333b", "address": "fa:16:3e:a5:4e:a7", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72964168-fd", "ovs_interfaceid": "72964168-fdd5-43bb-9940-86f4349e333b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 781.139301] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5c908f9c-7efb-4588-b6bf-70babb173767 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.143657] env[61907]: INFO nova.compute.manager [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Terminating instance [ 781.144946] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Acquired lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.145159] env[61907]: DEBUG nova.network.neutron [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Refreshing network info cache for port 72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.145806] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:4e:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16c6ea68-9b0e-4ac0-a484-7a9a40533017', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72964168-fdd5-43bb-9940-86f4349e333b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.153061] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating folder: Project (b57c1194e0f14ae498d99d302da346c5). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.153941] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243688, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065328} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.155311] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-894f7f19-d0ac-4417-801d-a75a730e9e4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.157328] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.162254] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfa88d2-a90f-443d-93ff-805a5a6c15cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.164736] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243687, 'name': CreateSnapshot_Task, 'duration_secs': 0.628531} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.166238] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 781.167443] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda82c37-c0fb-40ce-b606-73a2666fcd0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.187497] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] 3c9ee76a-464d-437e-9b73-862f1da24f7e/3c9ee76a-464d-437e-9b73-862f1da24f7e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.188804] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d9461ed-8b70-49db-b849-68ab0dec124b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.209236] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created folder: Project (b57c1194e0f14ae498d99d302da346c5) in parent group-v268168. [ 781.209462] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating folder: Instances. Parent ref: group-v268204. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 781.212591] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-493c4532-5051-4480-a261-a61fafaf26f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.218971] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 781.218971] env[61907]: value = "task-1243690" [ 781.218971] env[61907]: _type = "Task" [ 781.218971] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.225218] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created folder: Instances in parent group-v268204. [ 781.225218] env[61907]: DEBUG oslo.service.loopingcall [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.225218] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 781.225218] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62ee2de2-a66e-412e-8391-51d710e1ee7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.245464] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243690, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.246412] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.252150] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.252150] env[61907]: value = "task-1243692" [ 781.252150] env[61907]: _type = "Task" [ 781.252150] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.261971] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243692, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.350197] env[61907]: DEBUG nova.network.neutron [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [{"id": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "address": "fa:16:3e:0f:68:53", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05a00a60-c2", "ovs_interfaceid": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.657123] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e1f8f88b-7179-4e92-83cd-a1027ff9b0a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 781.665661] env[61907]: DEBUG nova.compute.manager [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 781.665661] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.665661] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7501a65e-88f8-4722-b843-79dbfec850d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.679915] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.679915] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28e6ea6c-3880-43e2-9cf6-8fd6c3536a54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.686734] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 781.686734] env[61907]: value = "task-1243693" [ 781.686734] env[61907]: _type = "Task" [ 781.686734] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.695010] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243693, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.723606] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 781.724496] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-61bc42bf-a5e7-4747-a633-0067d885ba91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.741205] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243690, 'name': ReconfigVM_Task, 'duration_secs': 0.37274} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.742872] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Reconfigured VM instance instance-00000034 to attach disk [datastore2] 3c9ee76a-464d-437e-9b73-862f1da24f7e/3c9ee76a-464d-437e-9b73-862f1da24f7e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.743277] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 781.743277] env[61907]: value = "task-1243694" [ 781.743277] env[61907]: _type = "Task" [ 781.743277] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.743464] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-097db109-0cd4-4e77-bb38-c95683c94a7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.754399] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243694, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.760362] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 781.760362] env[61907]: value = "task-1243695" [ 781.760362] env[61907]: _type = "Task" [ 781.760362] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.767133] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243692, 'name': CreateVM_Task, 'duration_secs': 0.338175} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.767627] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.768328] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.768502] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.768823] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 781.769063] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8adaf00c-8c40-4e95-87b0-abf7ddfbba0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.773328] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243695, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.777749] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 781.777749] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4924-6ea0-5ed8-a7e3-ef5aa6f256e9" [ 781.777749] env[61907]: _type = "Task" [ 781.777749] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.786062] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4924-6ea0-5ed8-a7e3-ef5aa6f256e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.852233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.919244] env[61907]: DEBUG nova.network.neutron [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Updated VIF entry in instance network info cache for port 72964168-fdd5-43bb-9940-86f4349e333b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 781.919609] env[61907]: DEBUG nova.network.neutron [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Updating instance_info_cache with network_info: [{"id": "72964168-fdd5-43bb-9940-86f4349e333b", "address": "fa:16:3e:a5:4e:a7", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap72964168-fd", "ovs_interfaceid": "72964168-fdd5-43bb-9940-86f4349e333b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.159178] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.196973] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243693, 'name': PowerOffVM_Task, 'duration_secs': 0.271758} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.197204] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 782.197434] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.197698] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56cd92be-e537-411a-87ef-89616519f9d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.254900] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243694, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.269967] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243695, 'name': Rename_Task, 'duration_secs': 0.193425} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.269967] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.270231] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b69ec6db-321b-4b1b-a479-babaa39b84dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.276456] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 782.276456] env[61907]: value = "task-1243697" [ 782.276456] env[61907]: _type = "Task" [ 782.276456] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.286836] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243697, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.290277] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4924-6ea0-5ed8-a7e3-ef5aa6f256e9, 'name': SearchDatastore_Task, 'duration_secs': 0.015612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.290559] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.290834] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.291054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.291206] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.291388] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.291643] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05a32c28-1b2b-4261-a6cf-8846eda587c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.300100] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.300233] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 782.301111] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d55877c-c010-4b01-a06f-6db1406cb8db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.307091] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 782.307091] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522d405f-d056-bd49-7e32-1692cae6b0bb" [ 782.307091] env[61907]: _type = "Task" [ 782.307091] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.314716] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522d405f-d056-bd49-7e32-1692cae6b0bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.422930] env[61907]: DEBUG oslo_concurrency.lockutils [req-6305a9fa-1f1f-4dc4-8de5-c867ffcf050d req-df9cebb5-9f20-4648-9654-309ea905ef43 service nova] Releasing lock "refresh_cache-1ef8594c-5d9a-491f-89f6-e5d7f35ce751" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.428537] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.428781] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.428966] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Deleting the datastore file [datastore1] 9209400d-d5b8-4b96-9022-f8a4c7a30471 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.429277] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-85fb5550-910d-4c9e-b871-8d31fe44183a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.437611] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for the task: (returnval){ [ 782.437611] env[61907]: value = "task-1243698" [ 782.437611] env[61907]: _type = "Task" [ 782.437611] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.446425] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243698, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.662728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 782.756565] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243694, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.786358] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243697, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.816822] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522d405f-d056-bd49-7e32-1692cae6b0bb, 'name': SearchDatastore_Task, 'duration_secs': 0.012536} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.817770] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0d7bb07-be1f-442b-815a-887953fe992e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.823118] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 782.823118] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52036731-559d-6806-3699-2ac2beb404ae" [ 782.823118] env[61907]: _type = "Task" [ 782.823118] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.831055] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52036731-559d-6806-3699-2ac2beb404ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.895036] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 782.895190] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ceb3f90-dfe3-4dda-9026-3c478b969cd1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.902211] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 782.902211] env[61907]: value = "task-1243699" [ 782.902211] env[61907]: _type = "Task" [ 782.902211] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.910041] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243699, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.946841] env[61907]: DEBUG oslo_vmware.api [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Task: {'id': task-1243698, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212639} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.947102] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.947290] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.947454] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.947629] env[61907]: INFO nova.compute.manager [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Took 1.29 seconds to destroy the instance on the hypervisor. [ 782.947868] env[61907]: DEBUG oslo.service.loopingcall [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.948077] env[61907]: DEBUG nova.compute.manager [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 782.948172] env[61907]: DEBUG nova.network.neutron [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 783.165914] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23526967-cd8f-4581-b9c5-1c270d385163 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.231888] env[61907]: DEBUG nova.compute.manager [req-279a30c8-265f-4440-81e6-1df58f4055ae req-fadd6839-b8d8-45e1-82b5-9e711931950e service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Received event network-vif-deleted-c7b1cf36-0ee3-4d59-b61c-1af1799938f5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 783.231888] env[61907]: INFO nova.compute.manager [req-279a30c8-265f-4440-81e6-1df58f4055ae req-fadd6839-b8d8-45e1-82b5-9e711931950e service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Neutron deleted interface c7b1cf36-0ee3-4d59-b61c-1af1799938f5; detaching it from the instance and deleting it from the info cache [ 783.231888] env[61907]: DEBUG nova.network.neutron [req-279a30c8-265f-4440-81e6-1df58f4055ae req-fadd6839-b8d8-45e1-82b5-9e711931950e service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.256024] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243694, 'name': CloneVM_Task, 'duration_secs': 1.489848} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.256024] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Created linked-clone VM from snapshot [ 783.256237] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb8462d-b1b1-4a00-8d86-4e47b25c8c00 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.263590] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Uploading image 33f4286c-8fff-4e4c-bb9b-2c0b2c3e1d03 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 783.273934] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 783.274186] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ae29f69d-3b6f-4dcc-b51c-8b1e26b56c47 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.281114] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 783.281114] env[61907]: value = "task-1243700" [ 783.281114] env[61907]: _type = "Task" [ 783.281114] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.286938] env[61907]: DEBUG oslo_vmware.api [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243697, 'name': PowerOnVM_Task, 'duration_secs': 0.520124} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.287479] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.287708] env[61907]: INFO nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Took 8.20 seconds to spawn the instance on the hypervisor. [ 783.287910] env[61907]: DEBUG nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 783.288698] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b01c2b-1d84-46f0-ad68-3e2d63c0fde1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.293924] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243700, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.333337] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52036731-559d-6806-3699-2ac2beb404ae, 'name': SearchDatastore_Task, 'duration_secs': 0.010592} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.334315] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.334663] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.334922] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-203831f9-544d-48f8-b844-4d82302dc487 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.342901] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 783.342901] env[61907]: value = "task-1243701" [ 783.342901] env[61907]: _type = "Task" [ 783.342901] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.351060] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.414438] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243699, 'name': PowerOffVM_Task, 'duration_secs': 0.226942} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.416440] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 783.416440] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6797836e-af25-479f-8f4f-beb38d660a1b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.433501] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f48dcfae-f390-4b05-93b2-1c052cbf07d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.466195] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.466549] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c60c4ea-4593-416d-bbd2-c5fb711c608d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.475093] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 783.475093] env[61907]: value = "task-1243702" [ 783.475093] env[61907]: _type = "Task" [ 783.475093] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.483645] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 783.483862] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.484121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.484340] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.484530] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.484821] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88f10d16-675b-4dee-aad3-affb3d76f7f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.493544] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.493732] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.494437] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9236cdb9-3476-4e58-820b-6917684bb42a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.499699] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 783.499699] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bc9773-1540-e2df-9e68-55c90ec05efb" [ 783.499699] env[61907]: _type = "Task" [ 783.499699] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.507017] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bc9773-1540-e2df-9e68-55c90ec05efb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.669162] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance da490943-511b-4776-8f16-4f51c3b055c4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 783.710493] env[61907]: DEBUG nova.network.neutron [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.734674] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-611a83d3-8d7b-4509-a575-42db4f1fb20b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.743779] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfd2a475-73a1-491c-87bb-c870a1e09360 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.769555] env[61907]: DEBUG nova.compute.manager [req-279a30c8-265f-4440-81e6-1df58f4055ae req-fadd6839-b8d8-45e1-82b5-9e711931950e service nova] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Detach interface failed, port_id=c7b1cf36-0ee3-4d59-b61c-1af1799938f5, reason: Instance 9209400d-d5b8-4b96-9022-f8a4c7a30471 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 783.790816] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243700, 'name': Destroy_Task, 'duration_secs': 0.374926} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.791143] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Destroyed the VM [ 783.791470] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 783.791739] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0374e5ad-5c75-4316-911b-bb37ad6f8af0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.798800] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 783.798800] env[61907]: value = "task-1243703" [ 783.798800] env[61907]: _type = "Task" [ 783.798800] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.811242] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243703, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.812509] env[61907]: INFO nova.compute.manager [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Took 31.87 seconds to build instance. [ 783.852522] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243701, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.010036] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bc9773-1540-e2df-9e68-55c90ec05efb, 'name': SearchDatastore_Task, 'duration_secs': 0.008922} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.010993] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f46bae27-8d97-404d-ab45-0a99e311af2a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.016481] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 784.016481] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520c2bfd-f38e-40c1-c45f-9cdb0b9c060f" [ 784.016481] env[61907]: _type = "Task" [ 784.016481] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.024242] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520c2bfd-f38e-40c1-c45f-9cdb0b9c060f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.172108] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 784.172534] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 784.172644] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 784.214040] env[61907]: INFO nova.compute.manager [-] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Took 1.27 seconds to deallocate network for instance. [ 784.311376] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243703, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.313918] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d6cbd88-ebcd-4952-80b2-5b7a515a0e5c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.222s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.355401] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586692} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.355719] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.355884] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.356145] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b3c2f723-0aa5-490e-890c-cffa9697dda6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.362470] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 784.362470] env[61907]: value = "task-1243704" [ 784.362470] env[61907]: _type = "Task" [ 784.362470] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.370892] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.484890] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcf164b2-5d9c-4197-8d36-805fa1b009f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.494731] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c62c134f-a26b-4e80-b849-e0cfe649fd4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.524963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.525239] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.525411] env[61907]: INFO nova.compute.manager [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Shelving [ 784.530930] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d34d4bc-39ba-4a82-9b38-db3865902d35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.541010] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520c2bfd-f38e-40c1-c45f-9cdb0b9c060f, 'name': SearchDatastore_Task, 'duration_secs': 0.039475} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.541320] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.541563] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 784.542736] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dbbcda-8cc3-4a32-af91-4f08ed43c98f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.546225] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a19ae6c-189d-4a32-9d6c-ea30f786c730 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.560012] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 784.562129] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 784.562129] env[61907]: value = "task-1243705" [ 784.562129] env[61907]: _type = "Task" [ 784.562129] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.569294] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.720113] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.811979] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243703, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.818590] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 784.872130] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06606} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.872385] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.873266] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d0f281-25d2-4a35-9df0-c6bd2c518630 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.896884] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.897651] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ada6c319-de84-4302-9ef0-d06b3075ab10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.917814] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 784.917814] env[61907]: value = "task-1243706" [ 784.917814] env[61907]: _type = "Task" [ 784.917814] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.925877] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243706, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.063445] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 785.076278] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473126} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.076771] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 785.077465] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e802ae-bad0-40d5-9498-260fa53d9165 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.105483] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.106075] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-211801c1-0253-43b3-94e7-c150494ee480 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.124563] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 785.124563] env[61907]: value = "task-1243707" [ 785.124563] env[61907]: _type = "Task" [ 785.124563] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.134025] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243707, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.310865] env[61907]: DEBUG oslo_vmware.api [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243703, 'name': RemoveSnapshot_Task, 'duration_secs': 1.234121} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.311161] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 785.339814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.429180] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243706, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.542710] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.543088] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f47d389f-2ca7-4dc8-83b0-7fc1d438cc25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.550267] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 785.550267] env[61907]: value = "task-1243708" [ 785.550267] env[61907]: _type = "Task" [ 785.550267] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.558457] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243708, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.572652] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 785.573229] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.573333] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.142s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.636264] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243707, 'name': ReconfigVM_Task, 'duration_secs': 0.435973} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.636680] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.637735] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a621b8c5-b586-4ba0-94af-4775e74a6555 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.665576] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a0c3a0f-f269-475c-882c-fa5eae7cdbbe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.681222] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 785.681222] env[61907]: value = "task-1243709" [ 785.681222] env[61907]: _type = "Task" [ 785.681222] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.690524] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243709, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.817141] env[61907]: WARNING nova.compute.manager [None req-823d497e-d235-4e37-a0fc-240b848f912c tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Image not found during snapshot: nova.exception.ImageNotFound: Image 33f4286c-8fff-4e4c-bb9b-2c0b2c3e1d03 could not be found. [ 785.929623] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243706, 'name': ReconfigVM_Task, 'duration_secs': 0.535281} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.929886] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.930503] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f44da57f-cdd3-4d09-b16d-b087ed28099c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.936281] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 785.936281] env[61907]: value = "task-1243710" [ 785.936281] env[61907]: _type = "Task" [ 785.936281] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.943094] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243710, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.060753] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243708, 'name': PowerOffVM_Task, 'duration_secs': 0.183767} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.061063] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 786.061867] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-059519ab-a352-49e4-b169-d23b85592b29 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.083457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83dd34a-97e1-4f96-b083-7ecd14f2bb3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.190774] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243709, 'name': ReconfigVM_Task, 'duration_secs': 0.18372} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.191105] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.191356] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ca2e92a-4d9e-4b70-8934-fa70f33e2b61 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.197354] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 786.197354] env[61907]: value = "task-1243711" [ 786.197354] env[61907]: _type = "Task" [ 786.197354] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.206947] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243711, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.267263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "a6a117b3-7cc4-43c5-8202-12514f695e48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.267889] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.267889] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.268145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.268245] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.272938] env[61907]: INFO nova.compute.manager [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Terminating instance [ 786.373020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b505aa49-3555-409d-b983-61c8792c92d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.380164] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4d19f8e-f60f-401b-91b5-7b9275276cec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.411535] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f11dd95-f15b-4775-a0b7-986361fca225 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.419013] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20a9db3-4d43-416b-a777-76cc964302d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.432889] env[61907]: DEBUG nova.compute.provider_tree [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.445807] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243710, 'name': Rename_Task, 'duration_secs': 0.152825} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.446413] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.446413] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8371afa2-f8fc-42ae-9530-570acc1f82ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.453032] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 786.453032] env[61907]: value = "task-1243712" [ 786.453032] env[61907]: _type = "Task" [ 786.453032] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.459831] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243712, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.593878] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 786.594273] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-728bf159-9493-42bb-a918-74776a37695c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.601491] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 786.601491] env[61907]: value = "task-1243713" [ 786.601491] env[61907]: _type = "Task" [ 786.601491] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.610255] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243713, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.707332] env[61907]: DEBUG oslo_vmware.api [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243711, 'name': PowerOnVM_Task, 'duration_secs': 0.399915} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.709045] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 786.710615] env[61907]: DEBUG nova.compute.manager [None req-4e883420-e16b-459c-b6e1-af2befbc113e tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 786.711552] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba0e545-020c-4287-8cb0-91d40a53abc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.778166] env[61907]: DEBUG nova.compute.manager [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 786.778400] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 786.779373] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86eccd88-9c25-43f4-89cf-5fd4d7eff281 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.786690] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 786.786932] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6ba8763c-9536-4490-bea7-4a94cb92e65e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.791917] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 786.791917] env[61907]: value = "task-1243714" [ 786.791917] env[61907]: _type = "Task" [ 786.791917] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.799848] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243714, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.935828] env[61907]: DEBUG nova.scheduler.client.report [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 786.964328] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243712, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.111922] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243713, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.308505] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243714, 'name': PowerOffVM_Task, 'duration_secs': 0.202488} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.308862] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 787.309151] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 787.309552] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d25ff9c-5504-47c7-8241-bbeb7dc80b51 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.373029] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 787.373029] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 787.373029] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleting the datastore file [datastore1] a6a117b3-7cc4-43c5-8202-12514f695e48 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 787.373356] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d7c1549e-9548-4e54-9aaf-b07748bfd346 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.379389] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 787.379389] env[61907]: value = "task-1243716" [ 787.379389] env[61907]: _type = "Task" [ 787.379389] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.387842] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.441068] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.868s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.441794] env[61907]: ERROR nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Traceback (most recent call last): [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.driver.spawn(context, instance, image_meta, [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] vm_ref = self.build_virtual_machine(instance, [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.441794] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] for vif in network_info: [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self._sync_wrapper(fn, *args, **kwargs) [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.wait() [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self[:] = self._gt.wait() [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self._exit_event.wait() [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] result = hub.switch() [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.442118] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return self.greenlet.switch() [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] result = function(*args, **kwargs) [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] return func(*args, **kwargs) [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise e [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] nwinfo = self.network_api.allocate_for_instance( [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] created_port_ids = self._update_ports_for_instance( [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] with excutils.save_and_reraise_exception(): [ 787.442429] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] self.force_reraise() [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise self.value [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] updated_port = self._update_port( [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] _ensure_no_port_binding_failure(port) [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] raise exception.PortBindingFailed(port_id=port['id']) [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] nova.exception.PortBindingFailed: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. [ 787.442722] env[61907]: ERROR nova.compute.manager [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] [ 787.442958] env[61907]: DEBUG nova.compute.utils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.443909] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.040s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.444157] env[61907]: DEBUG nova.objects.instance [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 787.451148] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Build of instance 215d0f35-5d19-4db4-996d-571e2970e19c was re-scheduled: Binding failed for port 30ec1c1b-b2e0-4890-a573-18042eb14e44, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 787.451648] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 787.451895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquiring lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.452077] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Acquired lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.452267] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.465728] env[61907]: DEBUG oslo_vmware.api [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243712, 'name': PowerOnVM_Task, 'duration_secs': 0.594863} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.466169] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.466802] env[61907]: INFO nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Took 9.87 seconds to spawn the instance on the hypervisor. [ 787.466802] env[61907]: DEBUG nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 787.467415] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef018b60-23fb-451c-a848-5a9e0dcc1137 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.612595] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243713, 'name': CreateSnapshot_Task, 'duration_secs': 0.678087} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.612789] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 787.613493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea279ca-1f7b-473f-8cb4-0d3a64c7709f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.889495] env[61907]: DEBUG oslo_vmware.api [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131072} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.889756] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 787.889943] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 787.890140] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.890341] env[61907]: INFO nova.compute.manager [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Took 1.11 seconds to destroy the instance on the hypervisor. [ 787.890604] env[61907]: DEBUG oslo.service.loopingcall [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.890792] env[61907]: DEBUG nova.compute.manager [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 787.890906] env[61907]: DEBUG nova.network.neutron [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.983884] env[61907]: INFO nova.compute.manager [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Took 32.24 seconds to build instance. [ 787.985909] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.079352] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.131231] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 788.131532] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d961d3a0-323e-4703-a8fc-ac2cb73ffcf0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.139903] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 788.139903] env[61907]: value = "task-1243717" [ 788.139903] env[61907]: _type = "Task" [ 788.139903] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.149625] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243717, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.186855] env[61907]: DEBUG nova.compute.manager [req-ed8fba78-f245-41dd-9f8e-4e97ea348791 req-81fca07f-5fc9-4bd1-8b35-b352a314f7ef service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Received event network-vif-deleted-8ed80238-7bda-433c-95f1-38554331b3de {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 788.186855] env[61907]: INFO nova.compute.manager [req-ed8fba78-f245-41dd-9f8e-4e97ea348791 req-81fca07f-5fc9-4bd1-8b35-b352a314f7ef service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Neutron deleted interface 8ed80238-7bda-433c-95f1-38554331b3de; detaching it from the instance and deleting it from the info cache [ 788.187117] env[61907]: DEBUG nova.network.neutron [req-ed8fba78-f245-41dd-9f8e-4e97ea348791 req-81fca07f-5fc9-4bd1-8b35-b352a314f7ef service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.336788] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.337243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.454466] env[61907]: DEBUG oslo_concurrency.lockutils [None req-078e4512-7b56-44aa-aaf6-baec99fe649f tempest-ServersAdmin275Test-720470488 tempest-ServersAdmin275Test-720470488-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.455599] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.297s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.488567] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09c3b2de-4bd5-4068-a954-a54360c7f1e2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.366s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.514044] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 788.514477] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 788.514733] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 788.582179] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Releasing lock "refresh_cache-215d0f35-5d19-4db4-996d-571e2970e19c" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.582462] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 788.582644] env[61907]: DEBUG nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 788.582811] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.606284] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.647079] env[61907]: DEBUG nova.network.neutron [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.651100] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243717, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.689856] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-08318178-bd65-43e8-81c5-c7559f7a24d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.698957] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd6ac41-5206-4457-91eb-5772d517aa4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.725048] env[61907]: DEBUG nova.compute.manager [req-ed8fba78-f245-41dd-9f8e-4e97ea348791 req-81fca07f-5fc9-4bd1-8b35-b352a314f7ef service nova] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Detach interface failed, port_id=8ed80238-7bda-433c-95f1-38554331b3de, reason: Instance a6a117b3-7cc4-43c5-8202-12514f695e48 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 788.994234] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 789.018862] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] There are 2 instances to clean {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 789.019360] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 67a7aef0-ed99-4bde-b2f0-33c36764a5d7] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 789.109438] env[61907]: DEBUG nova.network.neutron [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.152599] env[61907]: INFO nova.compute.manager [-] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Took 1.26 seconds to deallocate network for instance. [ 789.152954] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243717, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.403604] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5842bd60-9afa-4c05-99ba-c2bc45012354 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.411830] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5196d4-6a5c-41da-bab5-a335f0c13e62 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.446415] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2451fb76-d114-49ad-af62-2dfe040be27e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.454403] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89776062-786f-4ebe-9c71-de398774956e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.468615] env[61907]: DEBUG nova.compute.provider_tree [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.515636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.529759] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 48b683f0-9cfe-472c-97df-1c19c3b5e441] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 789.615432] env[61907]: INFO nova.compute.manager [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] [instance: 215d0f35-5d19-4db4-996d-571e2970e19c] Took 1.03 seconds to deallocate network for instance. [ 789.659358] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243717, 'name': CloneVM_Task, 'duration_secs': 1.286615} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.659646] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Created linked-clone VM from snapshot [ 789.660474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9431887-0535-4994-8c76-429a152ff8ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.666037] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.673978] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Uploading image b87d8f7b-8ebb-4e5f-8e92-7a6be207b773 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 789.701873] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 789.701873] env[61907]: value = "vm-268209" [ 789.701873] env[61907]: _type = "VirtualMachine" [ 789.701873] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 789.701873] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-91f33bbf-6db0-46c8-8f26-2b5c2ece3330 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.710402] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lease: (returnval){ [ 789.710402] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5287f33c-365f-09ef-92df-593ac2552586" [ 789.710402] env[61907]: _type = "HttpNfcLease" [ 789.710402] env[61907]: } obtained for exporting VM: (result){ [ 789.710402] env[61907]: value = "vm-268209" [ 789.710402] env[61907]: _type = "VirtualMachine" [ 789.710402] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 789.710681] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the lease: (returnval){ [ 789.710681] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5287f33c-365f-09ef-92df-593ac2552586" [ 789.710681] env[61907]: _type = "HttpNfcLease" [ 789.710681] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 789.718839] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 789.718839] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5287f33c-365f-09ef-92df-593ac2552586" [ 789.718839] env[61907]: _type = "HttpNfcLease" [ 789.718839] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 789.972036] env[61907]: DEBUG nova.scheduler.client.report [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 790.036024] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.036933] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances with incomplete migration {{(pid=61907) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 790.154879] env[61907]: INFO nova.compute.manager [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Rebuilding instance [ 790.203920] env[61907]: DEBUG nova.compute.manager [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 790.205036] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cda010b8-f934-4191-8041-0775c0f69feb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.220783] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 790.220783] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5287f33c-365f-09ef-92df-593ac2552586" [ 790.220783] env[61907]: _type = "HttpNfcLease" [ 790.220783] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 790.221148] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 790.221148] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5287f33c-365f-09ef-92df-593ac2552586" [ 790.221148] env[61907]: _type = "HttpNfcLease" [ 790.221148] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 790.221723] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aceaa917-06e2-4c03-a22d-7eef9291d174 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.229513] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 790.229690] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 790.382790] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-046bc634-fa74-4fd8-b80e-1b8add280967 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.477554] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.022s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.478261] env[61907]: ERROR nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Traceback (most recent call last): [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 2647, in _build_and_run_instance [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.driver.spawn(context, instance, image_meta, [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 542, in spawn [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self._vmops.spawn(context, instance, image_meta, injected_files, [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] vm_ref = self.build_virtual_machine(instance, [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] vif_infos = vmwarevif.get_vif_info(self._session, [ 790.478261] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] for vif in network_info: [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self._sync_wrapper(fn, *args, **kwargs) [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.wait() [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self[:] = self._gt.wait() [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self._exit_event.wait() [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] result = hub.switch() [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 790.478553] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return self.greenlet.switch() [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] result = function(*args, **kwargs) [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] return func(*args, **kwargs) [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 2017, in _allocate_network_async [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise e [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/compute/manager.py", line 1995, in _allocate_network_async [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] nwinfo = self.network_api.allocate_for_instance( [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] created_port_ids = self._update_ports_for_instance( [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] with excutils.save_and_reraise_exception(): [ 790.478996] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] self.force_reraise() [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise self.value [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] updated_port = self._update_port( [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] _ensure_no_port_binding_failure(port) [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] raise exception.PortBindingFailed(port_id=port['id']) [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] nova.exception.PortBindingFailed: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. [ 790.479314] env[61907]: ERROR nova.compute.manager [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] [ 790.479611] env[61907]: DEBUG nova.compute.utils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. {{(pid=61907) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 790.480391] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.847s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.482486] env[61907]: INFO nova.compute.claims [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.486369] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Build of instance 45c58b31-d0cc-4b36-b744-448a38e11266 was re-scheduled: Binding failed for port 362874eb-244f-4fa8-a357-bde8871e3075, please check neutron logs for more information. {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2486}} [ 790.486369] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Unplugging VIFs for instance {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3012}} [ 790.486369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquiring lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.486558] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Acquired lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.486600] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 790.540557] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 790.646851] env[61907]: INFO nova.scheduler.client.report [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Deleted allocations for instance 215d0f35-5d19-4db4-996d-571e2970e19c [ 791.016084] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.142384] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.156501] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c3d83c04-271c-4091-8ce2-a7f6da14fe89 tempest-ServersTestMultiNic-2060468103 tempest-ServersTestMultiNic-2060468103-project-member] Lock "215d0f35-5d19-4db4-996d-571e2970e19c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.504s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.222249] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 791.222652] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6005c6a1-1edb-4bbc-8c2b-f9d5673e864f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.230359] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 791.230359] env[61907]: value = "task-1243719" [ 791.230359] env[61907]: _type = "Task" [ 791.230359] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.241566] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.646630] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Releasing lock "refresh_cache-45c58b31-d0cc-4b36-b744-448a38e11266" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.646908] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61907) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3035}} [ 791.647110] env[61907]: DEBUG nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 791.647289] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 791.659143] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 791.678492] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.743598] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243719, 'name': PowerOffVM_Task, 'duration_secs': 0.235194} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.743757] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 791.744082] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 791.747935] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fee5860-2ece-4956-b4f0-cdc39ec12a7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.756515] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 791.758192] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28972a66-483e-4114-9158-88d0ee97c39e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.831726] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 791.832083] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 791.832325] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleting the datastore file [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 791.832643] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d14f4898-44b5-4e25-9abb-21a836145f82 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.836638] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3250a82-ba2b-4e9f-a8b2-44526b791468 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.846766] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0948689-5112-44ae-879b-fe4a0287cd04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.851889] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 791.851889] env[61907]: value = "task-1243721" [ 791.851889] env[61907]: _type = "Task" [ 791.851889] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.897810] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b5eb5a-1f63-48c3-8eb0-ed1f20dfeebc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.906187] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243721, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.913039] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419b4fe0-b6d3-48b5-9e72-951c5711845c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.935353] env[61907]: DEBUG nova.compute.provider_tree [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.043564] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 792.043807] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 792.043979] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 792.044157] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 792.044317] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 792.182943] env[61907]: DEBUG nova.network.neutron [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.188577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.362759] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243721, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158586} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.362926] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 792.363975] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 792.363975] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.438474] env[61907]: DEBUG nova.scheduler.client.report [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 792.547592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.687560] env[61907]: INFO nova.compute.manager [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] [instance: 45c58b31-d0cc-4b36-b744-448a38e11266] Took 1.04 seconds to deallocate network for instance. [ 792.944149] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.944700] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 792.947173] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.051s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.947391] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.949443] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.623s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.951170] env[61907]: INFO nova.compute.claims [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 792.985669] env[61907]: INFO nova.scheduler.client.report [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Deleted allocations for instance da7ddc92-f3e3-4b4d-a14b-13fd25a72314 [ 793.400132] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.400132] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.400132] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.400413] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.400413] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.400613] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.400978] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.401361] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.401643] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.401917] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.402235] env[61907]: DEBUG nova.virt.hardware [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.403406] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a403fffc-f37a-49ac-b6c8-57b76b4b7940 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.412253] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e404a6-ad8e-4afa-b1d6-dcee872c493c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.427038] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:4e:a7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16c6ea68-9b0e-4ac0-a484-7a9a40533017', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '72964168-fdd5-43bb-9940-86f4349e333b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.436169] env[61907]: DEBUG oslo.service.loopingcall [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.436430] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 793.436841] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4b33d9a0-55f3-4dbb-9891-27a14c252870 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.454801] env[61907]: DEBUG nova.compute.utils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.458946] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 793.459261] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 793.462903] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.462903] env[61907]: value = "task-1243722" [ 793.462903] env[61907]: _type = "Task" [ 793.462903] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.472559] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243722, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.497740] env[61907]: DEBUG oslo_concurrency.lockutils [None req-538fa060-d2bb-44eb-8ce1-8de5856bb393 tempest-ServersAdmin275Test-1763722295 tempest-ServersAdmin275Test-1763722295-project-member] Lock "da7ddc92-f3e3-4b4d-a14b-13fd25a72314" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.805s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.545913] env[61907]: DEBUG nova.policy [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9115523810b84ab3a4e0f822709a8eaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc8646a92534c53ba7b3335eb1ab9e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 793.735910] env[61907]: INFO nova.scheduler.client.report [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Deleted allocations for instance 45c58b31-d0cc-4b36-b744-448a38e11266 [ 793.964210] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 793.980253] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243722, 'name': CreateVM_Task, 'duration_secs': 0.36431} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.980453] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.981228] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.981361] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.981838] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 793.982166] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-719965f2-85ea-4fcb-a606-b85d048e6c5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.986669] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 793.986669] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cf7a53-c71d-720f-3939-ffdea3ca94fd" [ 793.986669] env[61907]: _type = "Task" [ 793.986669] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.999631] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cf7a53-c71d-720f-3939-ffdea3ca94fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.145946] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Successfully created port: 73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.246605] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3aac9b8f-d2cc-4bea-901b-e25be095481a tempest-ServersTestManualDisk-40215298 tempest-ServersTestManualDisk-40215298-project-member] Lock "45c58b31-d0cc-4b36-b744-448a38e11266" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.976s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.352918] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e26657af-ac81-4970-a7aa-746730e0c974 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.362758] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dd88d2-a474-4374-a0fc-3c2733e35af4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.399525] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173c3c17-7b27-4c03-bf59-481a1b6e563e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.407639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d7d9f9b-3631-4300-92d3-7506d192f3de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.423566] env[61907]: DEBUG nova.compute.provider_tree [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.497495] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cf7a53-c71d-720f-3939-ffdea3ca94fd, 'name': SearchDatastore_Task, 'duration_secs': 0.0103} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.498209] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.498454] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.498688] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.499432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.499432] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.499432] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d07ea86-a6c0-4dbc-9883-ee79e094fdfd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.508924] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.508995] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 794.509870] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d23416d-fe1a-40a9-9bd3-859dd7ece9e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.521035] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 794.521035] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f9b36e-14bf-13d9-509b-0c6b7ff21b08" [ 794.521035] env[61907]: _type = "Task" [ 794.521035] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.535643] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f9b36e-14bf-13d9-509b-0c6b7ff21b08, 'name': SearchDatastore_Task, 'duration_secs': 0.013204} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.536500] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ceafd2d1-aa8d-4730-a10d-67de2f4190f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.542683] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 794.542683] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525a3d10-74b2-cb93-80ec-697f7a55c2db" [ 794.542683] env[61907]: _type = "Task" [ 794.542683] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.551737] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525a3d10-74b2-cb93-80ec-697f7a55c2db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.750597] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 794.927176] env[61907]: DEBUG nova.scheduler.client.report [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 794.978684] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 795.010123] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.010539] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.010767] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.011059] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.011306] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.011521] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.011808] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.012047] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.012292] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.012541] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.012779] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.014077] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781b3a44-89bd-4b29-b907-eadc510942df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.022359] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f173b2b-6e72-400f-b34d-5a701dedcb48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.051978] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525a3d10-74b2-cb93-80ec-697f7a55c2db, 'name': SearchDatastore_Task, 'duration_secs': 0.008835} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.052518] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.052866] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 795.053188] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b26adc7-cd26-4505-98d8-40d03a5b8716 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.060423] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 795.060423] env[61907]: value = "task-1243723" [ 795.060423] env[61907]: _type = "Task" [ 795.060423] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.069078] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243723, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.278245] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.435024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.435024] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 795.436548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.092s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.438860] env[61907]: INFO nova.compute.claims [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.580997] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243723, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481818} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.582584] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.585018] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.585018] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7db4394-5d06-481e-b067-688b446198f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.589718] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 795.589718] env[61907]: value = "task-1243724" [ 795.589718] env[61907]: _type = "Task" [ 795.589718] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.946722] env[61907]: DEBUG nova.compute.utils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 795.951222] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 795.952279] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.033140] env[61907]: DEBUG nova.policy [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9115523810b84ab3a4e0f822709a8eaa', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccc8646a92534c53ba7b3335eb1ab9e4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.062325] env[61907]: DEBUG nova.compute.manager [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Received event network-vif-plugged-73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 796.062846] env[61907]: DEBUG oslo_concurrency.lockutils [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] Acquiring lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.063213] env[61907]: DEBUG oslo_concurrency.lockutils [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.063414] env[61907]: DEBUG oslo_concurrency.lockutils [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.063606] env[61907]: DEBUG nova.compute.manager [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] No waiting events found dispatching network-vif-plugged-73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 796.063812] env[61907]: WARNING nova.compute.manager [req-0e7c44a3-6cd5-48cd-ac3e-10c9e4629d5e req-586eb1a9-4ae3-4668-b340-a1f0994842f8 service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Received unexpected event network-vif-plugged-73705eb9-ea43-4aef-a8f6-68aa71f9582b for instance with vm_state building and task_state spawning. [ 796.103081] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243724, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075092} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.103733] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.105339] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b1c2d4f-99a9-4ba1-ab68-fc8aef770837 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.135269] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Reconfiguring VM instance instance-00000035 to attach disk [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.136308] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Successfully updated port: 73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.137756] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-516301f3-7a3e-4cfd-99b2-86464e8ca024 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.154742] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.155099] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.156505] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.165495] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 796.165495] env[61907]: value = "task-1243725" [ 796.165495] env[61907]: _type = "Task" [ 796.165495] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.174838] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243725, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.452703] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 796.642894] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Successfully created port: a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 796.678760] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243725, 'name': ReconfigVM_Task, 'duration_secs': 0.440084} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.678760] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Reconfigured VM instance instance-00000035 to attach disk [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751/1ef8594c-5d9a-491f-89f6-e5d7f35ce751.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.679808] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-201f1982-23a5-47b9-83a7-92b77febb513 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.691553] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 796.691553] env[61907]: value = "task-1243726" [ 796.691553] env[61907]: _type = "Task" [ 796.691553] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.704068] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243726, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.738891] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.850500] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d57aa6-ded2-458c-909c-14d9c0097ed1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.863430] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448a6fd0-26d5-4a9f-9f3b-3450025b28d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.906888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346c174b-1305-4204-9c0e-21f23b73e161 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.917363] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a969cf5-de18-4d63-8ca6-d95709f9867c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.932060] env[61907]: DEBUG nova.compute.provider_tree [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.111234] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Updating instance_info_cache with network_info: [{"id": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "address": "fa:16:3e:b7:7f:7e", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73705eb9-ea", "ovs_interfaceid": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.203327] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243726, 'name': Rename_Task, 'duration_secs': 0.190787} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.203665] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.203965] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6ce8522-bcd9-4102-98cd-93445ebf0c94 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.211157] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 797.211157] env[61907]: value = "task-1243727" [ 797.211157] env[61907]: _type = "Task" [ 797.211157] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.219876] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243727, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.436091] env[61907]: DEBUG nova.scheduler.client.report [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 797.465461] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 797.498478] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 797.498610] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 797.498664] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 797.499763] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 797.499763] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 797.499763] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 797.499763] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 797.499763] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 797.499933] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 797.500476] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 797.500476] env[61907]: DEBUG nova.virt.hardware [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.501790] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca32688b-916c-4200-a1ca-e4add73eb837 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.511995] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8783772f-03ac-46db-8700-d5374f29c3f3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.615880] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.616119] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Instance network_info: |[{"id": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "address": "fa:16:3e:b7:7f:7e", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73705eb9-ea", "ovs_interfaceid": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 797.616620] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b7:7f:7e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7f41333-42ee-47f3-936c-d6701ab786d2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73705eb9-ea43-4aef-a8f6-68aa71f9582b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 797.625534] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Creating folder: Project (ccc8646a92534c53ba7b3335eb1ab9e4). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.625922] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-417e367d-3c8b-47f2-9031-30c3f9c34cfc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.637869] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Created folder: Project (ccc8646a92534c53ba7b3335eb1ab9e4) in parent group-v268168. [ 797.637869] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Creating folder: Instances. Parent ref: group-v268211. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 797.638499] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b670044a-b210-4dfc-b4fb-bf02a2253cf7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.648488] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Created folder: Instances in parent group-v268211. [ 797.648842] env[61907]: DEBUG oslo.service.loopingcall [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 797.649127] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 797.649753] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0fa2ddee-4944-4f57-ad29-012e6c228269 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.672132] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 797.672132] env[61907]: value = "task-1243730" [ 797.672132] env[61907]: _type = "Task" [ 797.672132] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.681067] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243730, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.721512] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243727, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.943165] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.506s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.943929] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 797.951661] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.162s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.957644] env[61907]: INFO nova.compute.claims [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.165852] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 798.166992] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-861c12ad-2678-41b2-ab83-57264668a3f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.173314] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 798.173585] env[61907]: ERROR oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk due to incomplete transfer. [ 798.180434] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0c1f2c60-62f6-412b-9feb-60af3a639c5b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.188546] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243730, 'name': CreateVM_Task, 'duration_secs': 0.402142} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.189224] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 798.189520] env[61907]: DEBUG oslo_vmware.rw_handles [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/525cc441-22fa-21bd-ecb3-ba3dfba149ec/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 798.189711] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Uploaded image b87d8f7b-8ebb-4e5f-8e92-7a6be207b773 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 798.193914] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 798.193914] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.195111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.195111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.195111] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f7ce3dbe-5873-4a5e-9203-706f3557d5cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.196977] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-538a589b-d74c-41f3-94fd-88247d3b9ef0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.201761] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 798.201761] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524137d9-de44-51f1-0c5b-c62e23bc3702" [ 798.201761] env[61907]: _type = "Task" [ 798.201761] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.206176] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 798.206176] env[61907]: value = "task-1243731" [ 798.206176] env[61907]: _type = "Task" [ 798.206176] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.215656] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524137d9-de44-51f1-0c5b-c62e23bc3702, 'name': SearchDatastore_Task} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.222617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.222887] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 798.223139] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.223314] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.223498] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 798.223805] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243731, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.224040] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-626cf69a-fcc9-424d-9ae1-0b1c90b6adfb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.231360] env[61907]: DEBUG oslo_vmware.api [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243727, 'name': PowerOnVM_Task, 'duration_secs': 0.60945} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.232527] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 798.232775] env[61907]: DEBUG nova.compute.manager [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 798.233253] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 798.233443] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 798.234840] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d14b28b-eb24-45fb-8792-d198e478f4a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.237994] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0edbac2d-14d3-47ed-8859-9c26e4c12b14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.250018] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 798.250018] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52aa4601-69f2-ca33-858b-061706b3689a" [ 798.250018] env[61907]: _type = "Task" [ 798.250018] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.253354] env[61907]: DEBUG nova.compute.manager [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Received event network-changed-73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 798.253551] env[61907]: DEBUG nova.compute.manager [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Refreshing instance network info cache due to event network-changed-73705eb9-ea43-4aef-a8f6-68aa71f9582b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 798.253759] env[61907]: DEBUG oslo_concurrency.lockutils [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] Acquiring lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.253901] env[61907]: DEBUG oslo_concurrency.lockutils [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] Acquired lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.254073] env[61907]: DEBUG nova.network.neutron [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Refreshing network info cache for port 73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 798.264543] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52aa4601-69f2-ca33-858b-061706b3689a, 'name': SearchDatastore_Task, 'duration_secs': 0.008892} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.273379] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-93632a37-85f9-4367-98d8-2a90fbfc0cd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.278545] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 798.278545] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a027d-1f85-bf88-e652-be177945f483" [ 798.278545] env[61907]: _type = "Task" [ 798.278545] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.287440] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a027d-1f85-bf88-e652-be177945f483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.448535] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "41c680fb-5450-43f3-9acb-5218e57a7da9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.449160] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.466444] env[61907]: DEBUG nova.compute.utils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.470357] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 798.471027] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 798.578719] env[61907]: DEBUG nova.policy [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8682003dee14f25ba212739afd3d335', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6a88771a8f84196ae271dbd57590c15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 798.716481] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243731, 'name': Destroy_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.756569] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.759866] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Successfully updated port: a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 798.794031] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a027d-1f85-bf88-e652-be177945f483, 'name': SearchDatastore_Task, 'duration_secs': 0.009736} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.794031] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.794181] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5b20f363-dfad-4c2b-b757-4da4e5bd0b99/5b20f363-dfad-4c2b-b757-4da4e5bd0b99.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 798.794648] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4b856ea-b447-4c7e-9a86-2759cc809984 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.802957] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 798.802957] env[61907]: value = "task-1243732" [ 798.802957] env[61907]: _type = "Task" [ 798.802957] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.815784] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.933306] env[61907]: DEBUG nova.compute.manager [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Received event network-vif-plugged-a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 798.933561] env[61907]: DEBUG oslo_concurrency.lockutils [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] Acquiring lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.933784] env[61907]: DEBUG oslo_concurrency.lockutils [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.933961] env[61907]: DEBUG oslo_concurrency.lockutils [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.934171] env[61907]: DEBUG nova.compute.manager [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] No waiting events found dispatching network-vif-plugged-a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 798.934435] env[61907]: WARNING nova.compute.manager [req-206db8e5-489b-4324-944a-bb99312add67 req-bd96a55b-ad33-4dd5-90b6-cefd55c7b0fe service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Received unexpected event network-vif-plugged-a486b310-da28-4ae5-929b-a4c6ab192b31 for instance with vm_state building and task_state spawning. [ 798.978766] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 799.041267] env[61907]: DEBUG nova.network.neutron [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Updated VIF entry in instance network info cache for port 73705eb9-ea43-4aef-a8f6-68aa71f9582b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 799.041955] env[61907]: DEBUG nova.network.neutron [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Updating instance_info_cache with network_info: [{"id": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "address": "fa:16:3e:b7:7f:7e", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73705eb9-ea", "ovs_interfaceid": "73705eb9-ea43-4aef-a8f6-68aa71f9582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.152379] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "f028beaf-4185-40d7-a730-abe91209f9d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.152690] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.224543] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243731, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.264749] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.264909] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.265079] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 799.318797] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243732, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448999} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.319080] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5b20f363-dfad-4c2b-b757-4da4e5bd0b99/5b20f363-dfad-4c2b-b757-4da4e5bd0b99.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 799.320176] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.320176] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68ae9afe-9ad7-48a3-b90d-59a7d85139df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.326437] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 799.326437] env[61907]: value = "task-1243733" [ 799.326437] env[61907]: _type = "Task" [ 799.326437] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.337766] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243733, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.419122] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Successfully created port: da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.548847] env[61907]: DEBUG oslo_concurrency.lockutils [req-f951d8ff-06f6-4b03-b1bc-8235487c490c req-14c5bbc3-c2f4-4533-8795-ba72b9ded8ec service nova] Releasing lock "refresh_cache-5b20f363-dfad-4c2b-b757-4da4e5bd0b99" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.603438] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d2d29e-5cd2-48e1-9f57-c7028932f1c1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.613506] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27c038a5-7ce7-4771-8d5a-93287ddee223 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.651017] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b97828a-fa74-4ac1-9231-65491037618d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.661560] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f105f419-d54c-49fb-a7c0-2d0d6d524622 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.678652] env[61907]: DEBUG nova.compute.provider_tree [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.719740] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243731, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.807044] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.839727] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243733, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063271} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.840021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.840795] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06d97bd-8ab1-4065-a02e-46cce9b7d2de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.866725] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 5b20f363-dfad-4c2b-b757-4da4e5bd0b99/5b20f363-dfad-4c2b-b757-4da4e5bd0b99.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.867042] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72a65038-f419-4247-8eff-0b69ce6b7265 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.887201] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 799.887201] env[61907]: value = "task-1243734" [ 799.887201] env[61907]: _type = "Task" [ 799.887201] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.895625] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243734, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.992663] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 800.021190] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.021500] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.021665] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.021854] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.022022] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.022177] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.022454] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.022625] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.022801] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.022971] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.023169] env[61907]: DEBUG nova.virt.hardware [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.024059] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cae9efc-7af8-4b5c-ac12-c2dd38a9fd63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.030672] env[61907]: DEBUG nova.network.neutron [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Updating instance_info_cache with network_info: [{"id": "a486b310-da28-4ae5-929b-a4c6ab192b31", "address": "fa:16:3e:59:10:d9", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa486b310-da", "ovs_interfaceid": "a486b310-da28-4ae5-929b-a4c6ab192b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.034961] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd66fe19-5638-4cef-b7e4-01d6f03adfa9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.181956] env[61907]: DEBUG nova.scheduler.client.report [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 800.219669] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243731, 'name': Destroy_Task, 'duration_secs': 1.795395} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.219985] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Destroyed the VM [ 800.220263] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 800.220526] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-26b06e28-6461-461f-90ab-25d990e1236f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.228413] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 800.228413] env[61907]: value = "task-1243735" [ 800.228413] env[61907]: _type = "Task" [ 800.228413] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.235932] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243735, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.264476] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.264757] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.321701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.321701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.321701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.321701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.322038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.322753] env[61907]: INFO nova.compute.manager [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Terminating instance [ 800.398635] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243734, 'name': ReconfigVM_Task, 'duration_secs': 0.281473} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.398983] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 5b20f363-dfad-4c2b-b757-4da4e5bd0b99/5b20f363-dfad-4c2b-b757-4da4e5bd0b99.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.399687] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a7c6b5cd-85b9-4e98-a5ad-2721c6aec8f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.406349] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 800.406349] env[61907]: value = "task-1243736" [ 800.406349] env[61907]: _type = "Task" [ 800.406349] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.416308] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243736, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.535594] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.536013] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Instance network_info: |[{"id": "a486b310-da28-4ae5-929b-a4c6ab192b31", "address": "fa:16:3e:59:10:d9", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa486b310-da", "ovs_interfaceid": "a486b310-da28-4ae5-929b-a4c6ab192b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 800.536571] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:59:10:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7f41333-42ee-47f3-936c-d6701ab786d2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a486b310-da28-4ae5-929b-a4c6ab192b31', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 800.545418] env[61907]: DEBUG oslo.service.loopingcall [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 800.545755] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 800.546043] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f7fcd21-5ee1-4dfc-81ea-2f1be6b7047c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.568685] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 800.568685] env[61907]: value = "task-1243737" [ 800.568685] env[61907]: _type = "Task" [ 800.568685] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.576858] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243737, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.686614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.735s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.687292] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 800.690690] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.075s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.691528] env[61907]: INFO nova.compute.claims [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.741018] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243735, 'name': RemoveSnapshot_Task, 'duration_secs': 0.45821} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.741018] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 800.741018] env[61907]: DEBUG nova.compute.manager [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 800.741461] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a73f319-fdee-4d9b-b686-1c4d11ada670 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.828629] env[61907]: DEBUG nova.compute.manager [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 800.828832] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.829871] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8703e90-ffc3-4c2b-94e9-2a33c5a0a700 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.838363] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.838652] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-419321c9-c777-4893-87da-ad6e91c7da0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.845589] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 800.845589] env[61907]: value = "task-1243738" [ 800.845589] env[61907]: _type = "Task" [ 800.845589] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.860324] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243738, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.916205] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243736, 'name': Rename_Task, 'duration_secs': 0.161993} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.916511] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 800.916776] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-526e6477-f0cb-4ca6-9ba9-959637cddeb2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.923239] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 800.923239] env[61907]: value = "task-1243739" [ 800.923239] env[61907]: _type = "Task" [ 800.923239] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.934926] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.964050] env[61907]: DEBUG nova.compute.manager [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Received event network-changed-a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 800.964342] env[61907]: DEBUG nova.compute.manager [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Refreshing instance network info cache due to event network-changed-a486b310-da28-4ae5-929b-a4c6ab192b31. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 800.964618] env[61907]: DEBUG oslo_concurrency.lockutils [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] Acquiring lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.964815] env[61907]: DEBUG oslo_concurrency.lockutils [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] Acquired lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.965257] env[61907]: DEBUG nova.network.neutron [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Refreshing network info cache for port a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.081305] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243737, 'name': CreateVM_Task, 'duration_secs': 0.396163} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.082625] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 801.082958] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.083264] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.084092] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.084733] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.084733] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.085265] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3796e161-1d04-4cb4-9163-207cb8f63f68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.090453] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 801.090453] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276c281-3f79-f383-2197-3ceab6e0250c" [ 801.090453] env[61907]: _type = "Task" [ 801.090453] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.098331] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276c281-3f79-f383-2197-3ceab6e0250c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.198612] env[61907]: DEBUG nova.compute.utils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.206951] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 801.206951] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.256462] env[61907]: INFO nova.compute.manager [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Shelve offloading [ 801.265875] env[61907]: DEBUG nova.policy [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.360274] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243738, 'name': PowerOffVM_Task, 'duration_secs': 0.174168} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.360274] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.360274] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.361686] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6e66a65a-abed-4666-936a-7af79270b083 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.367551] env[61907]: DEBUG nova.compute.manager [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Received event network-vif-plugged-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 801.367826] env[61907]: DEBUG oslo_concurrency.lockutils [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] Acquiring lock "ce864af6-aef6-4044-be64-8440ba175438-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.368223] env[61907]: DEBUG oslo_concurrency.lockutils [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] Lock "ce864af6-aef6-4044-be64-8440ba175438-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.369046] env[61907]: DEBUG oslo_concurrency.lockutils [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] Lock "ce864af6-aef6-4044-be64-8440ba175438-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.369321] env[61907]: DEBUG nova.compute.manager [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] No waiting events found dispatching network-vif-plugged-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 801.369584] env[61907]: WARNING nova.compute.manager [req-dcecd6b6-8ec4-48da-98bb-2b43a823341b req-8a01a8a3-79ad-4842-8403-e90417a026f9 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Received unexpected event network-vif-plugged-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 for instance with vm_state building and task_state spawning. [ 801.426140] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.428304] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.428304] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleting the datastore file [datastore2] 1ef8594c-5d9a-491f-89f6-e5d7f35ce751 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.429733] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Successfully updated port: da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 801.440610] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42aa4bbe-3c87-4feb-baa6-64fac2ca49ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.446196] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.446447] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.447566] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.459934] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243739, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.461736] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 801.461736] env[61907]: value = "task-1243741" [ 801.461736] env[61907]: _type = "Task" [ 801.461736] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.473330] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243741, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.601923] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276c281-3f79-f383-2197-3ceab6e0250c, 'name': SearchDatastore_Task, 'duration_secs': 0.009399} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.602371] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.602963] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.602963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.607020] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.607020] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.607020] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33c51a99-cd84-4b8f-be76-2f5319e7f828 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.612878] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.613118] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.613873] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f9f359b-b3a6-40cb-a66b-f24484194e33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.619366] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 801.619366] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b25511-7123-c9f2-ac86-7fa75f03784c" [ 801.619366] env[61907]: _type = "Task" [ 801.619366] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.627798] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b25511-7123-c9f2-ac86-7fa75f03784c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.706931] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 801.727510] env[61907]: DEBUG nova.network.neutron [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Updated VIF entry in instance network info cache for port a486b310-da28-4ae5-929b-a4c6ab192b31. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 801.729465] env[61907]: DEBUG nova.network.neutron [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Updating instance_info_cache with network_info: [{"id": "a486b310-da28-4ae5-929b-a4c6ab192b31", "address": "fa:16:3e:59:10:d9", "network": {"id": "1d2cfe9c-7a53-42cb-863b-70685754554d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-326700873-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ccc8646a92534c53ba7b3335eb1ab9e4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7f41333-42ee-47f3-936c-d6701ab786d2", "external-id": "nsx-vlan-transportzone-674", "segmentation_id": 674, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa486b310-da", "ovs_interfaceid": "a486b310-da28-4ae5-929b-a4c6ab192b31", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.736058] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Successfully created port: ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.761944] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 801.762307] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5daea410-6980-4352-b70c-879706ea53da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.770659] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 801.770659] env[61907]: value = "task-1243742" [ 801.770659] env[61907]: _type = "Task" [ 801.770659] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.782160] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 801.782440] env[61907]: DEBUG nova.compute.manager [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 801.783496] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-299e4978-6214-4b21-aa08-87a305cdc80c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.789926] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.790137] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.790755] env[61907]: DEBUG nova.network.neutron [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 801.942881] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243739, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.973943] env[61907]: DEBUG oslo_vmware.api [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243741, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14882} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.974598] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.974598] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.974773] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.974970] env[61907]: INFO nova.compute.manager [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Took 1.15 seconds to destroy the instance on the hypervisor. [ 801.976049] env[61907]: DEBUG oslo.service.loopingcall [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.976049] env[61907]: DEBUG nova.compute.manager [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 801.976049] env[61907]: DEBUG nova.network.neutron [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 802.005602] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.132138] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b25511-7123-c9f2-ac86-7fa75f03784c, 'name': SearchDatastore_Task, 'duration_secs': 0.008957} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.132138] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b38fd639-119a-4a85-93b7-838a69d0664d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.144910] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 802.144910] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a58059-7567-c961-5df8-34f65e00e5df" [ 802.144910] env[61907]: _type = "Task" [ 802.144910] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.154907] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a58059-7567-c961-5df8-34f65e00e5df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.175566] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d5660d9-948d-417c-97e9-e0644508c658 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.186991] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f663cd-fcf9-49a7-820e-9cbc9cbd0d7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.228509] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cf64cbe-00bf-41ef-9860-131b9e73676b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.233068] env[61907]: DEBUG oslo_concurrency.lockutils [req-74f05072-7636-4926-8811-40613259d0a5 req-08f207bb-22ca-4238-9187-dbb9e26718fd service nova] Releasing lock "refresh_cache-08647e48-9e24-4d7c-a9fd-4066317a7250" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.239222] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f16937-9680-44d8-b5fb-3f4a15102f48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.244016] env[61907]: DEBUG nova.network.neutron [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Updating instance_info_cache with network_info: [{"id": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "address": "fa:16:3e:71:9c:f5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5f6c5a-0d", "ovs_interfaceid": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.258691] env[61907]: DEBUG nova.compute.provider_tree [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.436811] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243739, 'name': PowerOnVM_Task, 'duration_secs': 1.059718} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.437385] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 802.437634] env[61907]: INFO nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Took 7.46 seconds to spawn the instance on the hypervisor. [ 802.437828] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 802.438652] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0c8d54-fea0-4baa-9813-121f60c83ae1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.655850] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a58059-7567-c961-5df8-34f65e00e5df, 'name': SearchDatastore_Task, 'duration_secs': 0.010289} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.655850] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.656126] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 08647e48-9e24-4d7c-a9fd-4066317a7250/08647e48-9e24-4d7c-a9fd-4066317a7250.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 802.656404] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba49adf5-3021-4e4f-94c4-7aa6e9e4b37a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.664592] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 802.664592] env[61907]: value = "task-1243743" [ 802.664592] env[61907]: _type = "Task" [ 802.664592] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.676078] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.734636] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 802.747391] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.747743] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance network_info: |[{"id": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "address": "fa:16:3e:71:9c:f5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5f6c5a-0d", "ovs_interfaceid": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 802.748192] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:9c:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'da5f6c5a-0d2b-4fc6-8c90-1bfc00457236', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 802.759864] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating folder: Project (e6a88771a8f84196ae271dbd57590c15). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.760575] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04e74910-2595-4f37-b1d8-96e00fb28a2c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.767089] env[61907]: DEBUG nova.scheduler.client.report [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 802.778728] env[61907]: DEBUG nova.network.neutron [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updating instance_info_cache with network_info: [{"id": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "address": "fa:16:3e:d1:48:17", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1994ca95-09", "ovs_interfaceid": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.785373] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created folder: Project (e6a88771a8f84196ae271dbd57590c15) in parent group-v268168. [ 802.785373] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating folder: Instances. Parent ref: group-v268215. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 802.785373] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1811b218-89b9-4218-b351-f01deedb8208 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.788724] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.788991] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.789276] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.789513] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.789696] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.789865] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.790461] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.790691] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.790909] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.791441] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.791661] env[61907]: DEBUG nova.virt.hardware [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.792694] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57111c43-6781-4e66-bf97-a58d468592ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.802976] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ef1356-975c-4ce8-ba50-dca2304e05a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.809757] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created folder: Instances in parent group-v268215. [ 802.809999] env[61907]: DEBUG oslo.service.loopingcall [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 802.810641] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 802.810862] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bb37ecd-ff5c-4a83-8540-abab9f6a350a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.843241] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 802.843241] env[61907]: value = "task-1243746" [ 802.843241] env[61907]: _type = "Task" [ 802.843241] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.853750] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243746, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.944647] env[61907]: DEBUG nova.network.neutron [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.961596] env[61907]: INFO nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Took 40.36 seconds to build instance. [ 803.003053] env[61907]: DEBUG nova.compute.manager [req-f26d4662-8f52-4be3-a1b2-ade9b3119381 req-91604a5d-4ca6-4a6c-a499-142bd08c887c service nova] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Received event network-vif-deleted-72964168-fdd5-43bb-9940-86f4349e333b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 803.179289] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471062} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.179289] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 08647e48-9e24-4d7c-a9fd-4066317a7250/08647e48-9e24-4d7c-a9fd-4066317a7250.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 803.179289] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 803.179289] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-15da36b4-aa24-4e51-9c4d-09d83d5431e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.186021] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 803.186021] env[61907]: value = "task-1243747" [ 803.186021] env[61907]: _type = "Task" [ 803.186021] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.192223] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243747, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.277538] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.278083] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 803.281151] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.035s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.283207] env[61907]: INFO nova.compute.claims [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.290023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.355385] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243746, 'name': CreateVM_Task, 'duration_secs': 0.381863} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.355614] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 803.356377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.356594] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.357288] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 803.357605] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7907faf7-3825-4be9-b947-516e1b7ebbdf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.362825] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 803.362825] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52661588-9e20-9f07-dfc9-dceee8ae464f" [ 803.362825] env[61907]: _type = "Task" [ 803.362825] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.371279] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52661588-9e20-9f07-dfc9-dceee8ae464f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.404059] env[61907]: DEBUG nova.compute.manager [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Received event network-changed-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 803.404481] env[61907]: DEBUG nova.compute.manager [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Refreshing instance network info cache due to event network-changed-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 803.404884] env[61907]: DEBUG oslo_concurrency.lockutils [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] Acquiring lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.404884] env[61907]: DEBUG oslo_concurrency.lockutils [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] Acquired lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.405250] env[61907]: DEBUG nova.network.neutron [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Refreshing network info cache for port da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 803.450663] env[61907]: INFO nova.compute.manager [-] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Took 1.47 seconds to deallocate network for instance. [ 803.464330] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.730s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.695855] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243747, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074088} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.696499] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.697493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5c02ba-51bd-45c2-8ed2-80ef47f9a856 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.702128] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.702869] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ca128a-ee03-4394-b762-015d0d591f5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.732073] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Reconfiguring VM instance instance-00000037 to attach disk [datastore1] 08647e48-9e24-4d7c-a9fd-4066317a7250/08647e48-9e24-4d7c-a9fd-4066317a7250.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.735124] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b67bbf9-8482-48e6-853b-49f2c732d311 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.753924] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 803.754387] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fd404a95-5f97-465b-aee6-dede12a86a08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.761812] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 803.761812] env[61907]: value = "task-1243748" [ 803.761812] env[61907]: _type = "Task" [ 803.761812] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.770645] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243748, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.787833] env[61907]: DEBUG nova.compute.utils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.790286] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 803.790575] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.816755] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 803.817157] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 803.817273] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleting the datastore file [datastore2] 3c9ee76a-464d-437e-9b73-862f1da24f7e {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 803.817575] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-53d366b0-b51b-4554-8149-ee0bdd3b5751 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.825632] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 803.825632] env[61907]: value = "task-1243750" [ 803.825632] env[61907]: _type = "Task" [ 803.825632] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.836554] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243750, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.874229] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52661588-9e20-9f07-dfc9-dceee8ae464f, 'name': SearchDatastore_Task, 'duration_secs': 0.008933} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.877437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.877437] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 803.877437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.877437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.877437] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 803.877614] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f29b9a8e-d3e2-47ea-88f8-a28bc542261d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.893575] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 803.893792] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 803.894588] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3649bc-17d9-4c25-bef2-e436fc9a8b4f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.900488] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 803.900488] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f1368a-d532-6ded-793f-880908b05a9d" [ 803.900488] env[61907]: _type = "Task" [ 803.900488] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.912691] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f1368a-d532-6ded-793f-880908b05a9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.956594] env[61907]: DEBUG nova.policy [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a850b5f34ac46daa2d7009ef04c95cd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '35661a1f7c784aa8938288ea955eae52', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.961083] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.968126] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 804.011100] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Successfully updated port: ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 804.274241] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243748, 'name': ReconfigVM_Task, 'duration_secs': 0.32412} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.275081] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Reconfigured VM instance instance-00000037 to attach disk [datastore1] 08647e48-9e24-4d7c-a9fd-4066317a7250/08647e48-9e24-4d7c-a9fd-4066317a7250.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.275171] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8cac08ca-1d8e-4d43-9cf2-b243ccdf01a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.281417] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 804.281417] env[61907]: value = "task-1243751" [ 804.281417] env[61907]: _type = "Task" [ 804.281417] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.292050] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243751, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.297797] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 804.304704] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Successfully created port: 369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.343634] env[61907]: DEBUG oslo_vmware.api [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243750, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168827} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.344033] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 804.344257] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 804.344444] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.372697] env[61907]: INFO nova.scheduler.client.report [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance 3c9ee76a-464d-437e-9b73-862f1da24f7e [ 804.416890] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f1368a-d532-6ded-793f-880908b05a9d, 'name': SearchDatastore_Task, 'duration_secs': 0.00994} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.417821] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae4ea8e4-fe4d-4132-b400-bb8890669cc6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.425977] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 804.425977] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bdaa8d-9e0d-4d57-716f-dc3446f858f8" [ 804.425977] env[61907]: _type = "Task" [ 804.425977] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.434562] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bdaa8d-9e0d-4d57-716f-dc3446f858f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.487478] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.515852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.515953] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.516124] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.617586] env[61907]: DEBUG nova.network.neutron [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Updated VIF entry in instance network info cache for port da5f6c5a-0d2b-4fc6-8c90-1bfc00457236. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 804.617586] env[61907]: DEBUG nova.network.neutron [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Updating instance_info_cache with network_info: [{"id": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "address": "fa:16:3e:71:9c:f5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapda5f6c5a-0d", "ovs_interfaceid": "da5f6c5a-0d2b-4fc6-8c90-1bfc00457236", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.718518] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9100432b-1012-49c6-a481-2b85673be913 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.727936] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c06b06e-463c-49f2-b600-334528fd8989 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.767256] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d5654a-9cca-4421-bee0-b7cfc91aa373 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.776944] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab0a9bd-c994-4e68-a321-20a03385b07b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.796140] env[61907]: DEBUG nova.compute.provider_tree [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.802727] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243751, 'name': Rename_Task, 'duration_secs': 0.139335} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.803580] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.803580] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-317059f7-f8df-43f8-b440-d2d51bf49bf6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.819020] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 804.819020] env[61907]: value = "task-1243752" [ 804.819020] env[61907]: _type = "Task" [ 804.819020] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.824819] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243752, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.881442] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.944759] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bdaa8d-9e0d-4d57-716f-dc3446f858f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008947} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.944759] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.944896] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] ce864af6-aef6-4044-be64-8440ba175438/ce864af6-aef6-4044-be64-8440ba175438.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 804.946390] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-59ef3401-3a8b-4fee-97f5-c07f214385ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.955462] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 804.955462] env[61907]: value = "task-1243753" [ 804.955462] env[61907]: _type = "Task" [ 804.955462] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.964937] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243753, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.074781] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.099385] env[61907]: DEBUG nova.compute.manager [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-plugged-ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 805.099709] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.100024] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.100613] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.100613] env[61907]: DEBUG nova.compute.manager [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] No waiting events found dispatching network-vif-plugged-ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 805.101118] env[61907]: WARNING nova.compute.manager [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received unexpected event network-vif-plugged-ce4e1e2e-3671-4aba-a760-b96db6f21180 for instance with vm_state building and task_state spawning. [ 805.101118] env[61907]: DEBUG nova.compute.manager [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-changed-ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 805.101259] env[61907]: DEBUG nova.compute.manager [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing instance network info cache due to event network-changed-ce4e1e2e-3671-4aba-a760-b96db6f21180. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 805.101547] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.120449] env[61907]: DEBUG oslo_concurrency.lockutils [req-6e33bda8-5866-4463-b5da-40502d039da1 req-785b3785-dbff-46c0-ae35-e989b0607015 service nova] Releasing lock "refresh_cache-ce864af6-aef6-4044-be64-8440ba175438" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.299483] env[61907]: DEBUG nova.scheduler.client.report [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 805.311393] env[61907]: DEBUG nova.network.neutron [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.316985] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 805.328847] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243752, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.351037] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.351541] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.351743] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.351999] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.352182] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.352410] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.352666] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.352860] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.353089] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.353351] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.353633] env[61907]: DEBUG nova.virt.hardware [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.354881] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e350a0-697f-47f7-af43-72d44a46e49b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.366532] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4015160-3f08-4bdf-9b9f-0bf54e436b42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.464919] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243753, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.535294] env[61907]: DEBUG nova.compute.manager [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received event network-vif-unplugged-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 805.535529] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.535738] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.535912] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.536306] env[61907]: DEBUG nova.compute.manager [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] No waiting events found dispatching network-vif-unplugged-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 805.536459] env[61907]: DEBUG nova.compute.manager [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received event network-vif-unplugged-1994ca95-0956-444e-8bb0-cd6a773e7ede for instance with task_state deleting. {{(pid=61907) _process_instance_event /opt/stack/nova/nova/compute/manager.py:11252}} [ 805.536641] env[61907]: DEBUG nova.compute.manager [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Received event network-changed-1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 805.536803] env[61907]: DEBUG nova.compute.manager [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Refreshing instance network info cache due to event network-changed-1994ca95-0956-444e-8bb0-cd6a773e7ede. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 805.536993] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Acquiring lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.537152] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Acquired lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.537315] env[61907]: DEBUG nova.network.neutron [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Refreshing network info cache for port 1994ca95-0956-444e-8bb0-cd6a773e7ede {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.605515] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.807531] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.808173] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 805.811905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.092s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.812167] env[61907]: DEBUG nova.objects.instance [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lazy-loading 'resources' on Instance uuid 9209400d-d5b8-4b96-9022-f8a4c7a30471 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.818631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.818790] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Instance network_info: |[{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 805.819174] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.819359] env[61907]: DEBUG nova.network.neutron [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing network info cache for port ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.820394] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:65:d7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce4e1e2e-3671-4aba-a760-b96db6f21180', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.832335] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Creating folder: Project (6df6b1cd82e24d2f8aa1812575f03f8a). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.838496] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-31cbc862-1464-4892-8048-20df6c23efe9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.848418] env[61907]: DEBUG oslo_vmware.api [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243752, 'name': PowerOnVM_Task, 'duration_secs': 0.657277} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.848755] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.848984] env[61907]: INFO nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Took 8.38 seconds to spawn the instance on the hypervisor. [ 805.849410] env[61907]: DEBUG nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 805.850460] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7f7dc9-b4b9-4299-af2c-28b7b9f02f8c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.857340] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Created folder: Project (6df6b1cd82e24d2f8aa1812575f03f8a) in parent group-v268168. [ 805.857340] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Creating folder: Instances. Parent ref: group-v268218. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.857848] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a3e8f214-9d78-497c-808a-3396634a6e39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.872802] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Created folder: Instances in parent group-v268218. [ 805.872943] env[61907]: DEBUG oslo.service.loopingcall [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.873134] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.873350] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b7f4f9a9-08aa-473e-b590-ea6f90a41d10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.895824] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.895824] env[61907]: value = "task-1243756" [ 805.895824] env[61907]: _type = "Task" [ 805.895824] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.903182] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243756, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.969504] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243753, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.522555} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.969504] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] ce864af6-aef6-4044-be64-8440ba175438/ce864af6-aef6-4044-be64-8440ba175438.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 805.969504] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 805.969504] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-826cceac-963f-4baf-9e9a-0ea3a59db188 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.976021] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 805.976021] env[61907]: value = "task-1243757" [ 805.976021] env[61907]: _type = "Task" [ 805.976021] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.984283] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243757, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.075829] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Successfully updated port: 369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 806.288215] env[61907]: DEBUG nova.network.neutron [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updated VIF entry in instance network info cache for port 1994ca95-0956-444e-8bb0-cd6a773e7ede. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.288652] env[61907]: DEBUG nova.network.neutron [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updating instance_info_cache with network_info: [{"id": "1994ca95-0956-444e-8bb0-cd6a773e7ede", "address": "fa:16:3e:d1:48:17", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": null, "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1994ca95-09", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.318644] env[61907]: DEBUG nova.compute.utils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 806.320900] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 806.321084] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 806.384038] env[61907]: INFO nova.compute.manager [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Took 32.08 seconds to build instance. [ 806.390926] env[61907]: DEBUG nova.policy [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fad5df91213b46d897316643f915087f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bfa11d39bdf14738a34f3534a44a9851', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 806.398857] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "aec01156-431b-447e-88ef-cb8d254c6341" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.399139] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.412148] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243756, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.489309] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243757, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.133598} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.489622] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 806.490405] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98ff6b7-ef5a-4e59-bd83-065eae964d84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.513653] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Reconfiguring VM instance instance-00000038 to attach disk [datastore1] ce864af6-aef6-4044-be64-8440ba175438/ce864af6-aef6-4044-be64-8440ba175438.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 806.516729] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b90c8cf9-4a3a-4af8-a8a0-03f524aba220 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.541311] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 806.541311] env[61907]: value = "task-1243758" [ 806.541311] env[61907]: _type = "Task" [ 806.541311] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.553835] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243758, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.579031] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.586537] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquired lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.586537] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 806.691015] env[61907]: DEBUG nova.network.neutron [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updated VIF entry in instance network info cache for port ce4e1e2e-3671-4aba-a760-b96db6f21180. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 806.691015] env[61907]: DEBUG nova.network.neutron [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.775278] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Successfully created port: 30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 806.779572] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9195f52f-1d86-4462-a8c9-0a5b28d4d200 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.787444] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-035fc3c2-3207-43b8-b267-d1d8d398f4ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.791795] env[61907]: DEBUG oslo_concurrency.lockutils [req-923d3d2f-baac-4cf1-9a75-c3a611d7a926 req-005d08f3-9b1e-4030-bb60-337b4c89af19 service nova] Releasing lock "refresh_cache-3c9ee76a-464d-437e-9b73-862f1da24f7e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.826298] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 806.831014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70d4c05-a203-422f-9938-f7b96fedfffc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.840244] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6875e6a-ae84-4ff4-8edb-b555da383676 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.855318] env[61907]: DEBUG nova.compute.provider_tree [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.890126] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be0a885f-7196-4e92-9305-b999c5053a15 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.130s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.907502] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243756, 'name': CreateVM_Task, 'duration_secs': 0.696948} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.907670] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.908375] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.908577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.908887] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 806.909159] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be047a86-ba22-44d2-a9bb-211648f2810f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.914064] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 806.914064] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3300-2247-5f63-1c31-8a6d83a90eaa" [ 806.914064] env[61907]: _type = "Task" [ 806.914064] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.922388] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3300-2247-5f63-1c31-8a6d83a90eaa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.052130] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243758, 'name': ReconfigVM_Task, 'duration_secs': 0.289315} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.055000] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Reconfigured VM instance instance-00000038 to attach disk [datastore1] ce864af6-aef6-4044-be64-8440ba175438/ce864af6-aef6-4044-be64-8440ba175438.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 807.055000] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7a10a369-077e-4a0d-bbac-75e58f785460 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.061548] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 807.061548] env[61907]: value = "task-1243759" [ 807.061548] env[61907]: _type = "Task" [ 807.061548] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.069202] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243759, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.153820] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.193759] env[61907]: DEBUG oslo_concurrency.lockutils [req-50b588b5-45e6-4389-a616-c18234bdad17 req-198d72a0-c37f-49cd-a98d-0a1983414080 service nova] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.234856] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.235191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.235424] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.235615] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.235789] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.240302] env[61907]: INFO nova.compute.manager [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Terminating instance [ 807.337039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "08647e48-9e24-4d7c-a9fd-4066317a7250" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.337305] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.337513] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.337740] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.337952] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.340158] env[61907]: DEBUG nova.network.neutron [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Updating instance_info_cache with network_info: [{"id": "369b29da-86e9-48a7-8cae-7627cd00ce30", "address": "fa:16:3e:ff:5a:77", "network": {"id": "2e53f126-57eb-40d4-8e13-fb599539212f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1889709567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35661a1f7c784aa8938288ea955eae52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap369b29da-86", "ovs_interfaceid": "369b29da-86e9-48a7-8cae-7627cd00ce30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.341548] env[61907]: INFO nova.compute.manager [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Terminating instance [ 807.361672] env[61907]: DEBUG nova.scheduler.client.report [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 807.393039] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 807.425480] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3300-2247-5f63-1c31-8a6d83a90eaa, 'name': SearchDatastore_Task, 'duration_secs': 0.04551} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.425480] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.425785] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.425905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.426057] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.426289] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.426775] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5d2571ec-0174-47b4-9810-712871e0333a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.444469] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.444665] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 807.445385] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f874bd-71ad-4673-8939-e6304f4f002a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.451050] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 807.451050] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215bb6e-51b1-75e2-521a-3dd953de2352" [ 807.451050] env[61907]: _type = "Task" [ 807.451050] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.458656] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215bb6e-51b1-75e2-521a-3dd953de2352, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.563547] env[61907]: DEBUG nova.compute.manager [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Received event network-vif-plugged-369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 807.563779] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Acquiring lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.563993] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.564184] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.564359] env[61907]: DEBUG nova.compute.manager [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] No waiting events found dispatching network-vif-plugged-369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 807.564532] env[61907]: WARNING nova.compute.manager [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Received unexpected event network-vif-plugged-369b29da-86e9-48a7-8cae-7627cd00ce30 for instance with vm_state building and task_state spawning. [ 807.564725] env[61907]: DEBUG nova.compute.manager [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Received event network-changed-369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 807.564907] env[61907]: DEBUG nova.compute.manager [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Refreshing instance network info cache due to event network-changed-369b29da-86e9-48a7-8cae-7627cd00ce30. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 807.565094] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Acquiring lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.573957] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243759, 'name': Rename_Task, 'duration_secs': 0.168784} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.574217] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 807.574448] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-18aa1f35-eb1e-4c8a-a4f0-3cb1b3692983 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.582768] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 807.582768] env[61907]: value = "task-1243760" [ 807.582768] env[61907]: _type = "Task" [ 807.582768] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.590524] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243760, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.744702] env[61907]: DEBUG nova.compute.manager [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.745113] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.746396] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f012a09-b7b1-4be2-aa45-3731416c76bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.754453] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.754717] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b99eb2a-bda7-4eb0-af53-0ee20372de70 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.761053] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 807.761053] env[61907]: value = "task-1243761" [ 807.761053] env[61907]: _type = "Task" [ 807.761053] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.770832] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243761, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.843432] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 807.846356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Releasing lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.846936] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Instance network_info: |[{"id": "369b29da-86e9-48a7-8cae-7627cd00ce30", "address": "fa:16:3e:ff:5a:77", "network": {"id": "2e53f126-57eb-40d4-8e13-fb599539212f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1889709567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35661a1f7c784aa8938288ea955eae52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap369b29da-86", "ovs_interfaceid": "369b29da-86e9-48a7-8cae-7627cd00ce30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 807.847418] env[61907]: DEBUG nova.compute.manager [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 807.847788] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.848059] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Acquired lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.848330] env[61907]: DEBUG nova.network.neutron [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Refreshing network info cache for port 369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 807.849783] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ff:5a:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '304be4f7-4e36-4468-9ef4-e457341cef18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '369b29da-86e9-48a7-8cae-7627cd00ce30', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.857904] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Creating folder: Project (35661a1f7c784aa8938288ea955eae52). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.858786] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a703bc0f-5481-4508-8204-75638640a559 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.865910] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4799749-bc20-45d8-8a82-af9e8ea8050b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.868458] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.057s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.871148] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.531s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.872812] env[61907]: INFO nova.compute.claims [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.878752] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.879120] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.879241] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.879383] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.879675] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.879883] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.880056] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.880382] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.880632] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.880847] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.881099] env[61907]: DEBUG nova.virt.hardware [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.885607] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4df56e0-8a40-45ad-a12f-03386e065a39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.889923] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.890217] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-903a12b1-904b-49d7-a3cb-027cc4901ef9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.894444] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Created folder: Project (35661a1f7c784aa8938288ea955eae52) in parent group-v268168. [ 807.894537] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Creating folder: Instances. Parent ref: group-v268221. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 807.895159] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea5324a4-82ec-4bb8-b9cf-9f5524ac3ccb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.898531] env[61907]: INFO nova.scheduler.client.report [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Deleted allocations for instance 9209400d-d5b8-4b96-9022-f8a4c7a30471 [ 807.909907] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f48f29-342e-4caf-b4d4-20a754c035d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.914795] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 807.914795] env[61907]: value = "task-1243763" [ 807.914795] env[61907]: _type = "Task" [ 807.914795] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.926280] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Created folder: Instances in parent group-v268221. [ 807.926592] env[61907]: DEBUG oslo.service.loopingcall [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.936239] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 807.942026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.942026] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-460fbc18-119a-4039-ae5a-bd8224a57f6c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.959962] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243763, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.970099] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215bb6e-51b1-75e2-521a-3dd953de2352, 'name': SearchDatastore_Task, 'duration_secs': 0.026717} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.972797] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.972797] env[61907]: value = "task-1243765" [ 807.972797] env[61907]: _type = "Task" [ 807.972797] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.973597] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efa6aa2b-a87f-407e-8f64-e0296be2ce6e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.981975] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 807.981975] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529db86b-b990-5337-7147-cd513605fa2f" [ 807.981975] env[61907]: _type = "Task" [ 807.981975] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.985841] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243765, 'name': CreateVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.998624] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529db86b-b990-5337-7147-cd513605fa2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009393} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.999016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.999326] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 52404147-0375-4f9c-9e1e-4a52001df7cb/52404147-0375-4f9c-9e1e-4a52001df7cb.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 807.999619] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-883658d5-7775-44d9-a551-91bb56633161 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.008592] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 808.008592] env[61907]: value = "task-1243766" [ 808.008592] env[61907]: _type = "Task" [ 808.008592] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.017659] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243766, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.096716] env[61907]: DEBUG oslo_vmware.api [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243760, 'name': PowerOnVM_Task, 'duration_secs': 0.446573} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.097048] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 808.097386] env[61907]: INFO nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Took 8.10 seconds to spawn the instance on the hypervisor. [ 808.097497] env[61907]: DEBUG nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 808.098329] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa34cf71-fc5c-4ad0-bc7c-fa035fe5064e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.276170] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243761, 'name': PowerOffVM_Task, 'duration_secs': 0.188289} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.276675] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.276983] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.277467] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e0a17d40-02c6-44ae-b3e8-368ca1936b69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.383529] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.384064] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.384166] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleting the datastore file [datastore2] 5b20f363-dfad-4c2b-b757-4da4e5bd0b99 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.387672] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a103221-5d73-44b2-af23-696698806101 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.397086] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 808.397086] env[61907]: value = "task-1243768" [ 808.397086] env[61907]: _type = "Task" [ 808.397086] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.408510] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243768, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.419668] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7976d913-1462-4722-a36b-4c696d0ec750 tempest-ServerMetadataTestJSON-1000286231 tempest-ServerMetadataTestJSON-1000286231-project-member] Lock "9209400d-d5b8-4b96-9022-f8a4c7a30471" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.286s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.434429] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243763, 'name': PowerOffVM_Task, 'duration_secs': 0.197223} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.434821] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.435067] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.435379] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8bc22d08-fadc-461a-85b4-6bf66d3d6665 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.491599] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243765, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.516660] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.516953] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.517194] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleting the datastore file [datastore1] 08647e48-9e24-4d7c-a9fd-4066317a7250 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.517903] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b86a9ebd-750b-4a83-9ef7-43b556e0a238 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.524242] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243766, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513186} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.525557] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 52404147-0375-4f9c-9e1e-4a52001df7cb/52404147-0375-4f9c-9e1e-4a52001df7cb.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 808.525817] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 808.526155] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for the task: (returnval){ [ 808.526155] env[61907]: value = "task-1243770" [ 808.526155] env[61907]: _type = "Task" [ 808.526155] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.526397] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c44f4fdb-5dbe-4f53-971d-45f7519ce7d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.537042] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.538623] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Successfully updated port: 30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 808.540393] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 808.540393] env[61907]: value = "task-1243771" [ 808.540393] env[61907]: _type = "Task" [ 808.540393] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.554357] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243771, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.584421] env[61907]: DEBUG nova.compute.manager [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Received event network-vif-plugged-30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 808.584674] env[61907]: DEBUG oslo_concurrency.lockutils [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] Acquiring lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.584896] env[61907]: DEBUG oslo_concurrency.lockutils [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.585278] env[61907]: DEBUG oslo_concurrency.lockutils [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.585517] env[61907]: DEBUG nova.compute.manager [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] No waiting events found dispatching network-vif-plugged-30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 808.585703] env[61907]: WARNING nova.compute.manager [req-71bca44f-6f76-411e-9e6e-669bb88a86ff req-881c3027-7071-4b5b-bb11-8b525f1c7697 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Received unexpected event network-vif-plugged-30bd832d-f042-4736-8dea-152ff21d9739 for instance with vm_state building and task_state spawning. [ 808.618617] env[61907]: INFO nova.compute.manager [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Took 33.30 seconds to build instance. [ 808.801307] env[61907]: DEBUG nova.network.neutron [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Updated VIF entry in instance network info cache for port 369b29da-86e9-48a7-8cae-7627cd00ce30. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 808.802230] env[61907]: DEBUG nova.network.neutron [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Updating instance_info_cache with network_info: [{"id": "369b29da-86e9-48a7-8cae-7627cd00ce30", "address": "fa:16:3e:ff:5a:77", "network": {"id": "2e53f126-57eb-40d4-8e13-fb599539212f", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1889709567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "35661a1f7c784aa8938288ea955eae52", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "304be4f7-4e36-4468-9ef4-e457341cef18", "external-id": "nsx-vlan-transportzone-911", "segmentation_id": 911, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap369b29da-86", "ovs_interfaceid": "369b29da-86e9-48a7-8cae-7627cd00ce30", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.911402] env[61907]: DEBUG oslo_vmware.api [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243768, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225622} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.911697] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.911883] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.912079] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.912262] env[61907]: INFO nova.compute.manager [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Took 1.17 seconds to destroy the instance on the hypervisor. [ 808.915014] env[61907]: DEBUG oslo.service.loopingcall [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.915014] env[61907]: DEBUG nova.compute.manager [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 808.915014] env[61907]: DEBUG nova.network.neutron [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.987952] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243765, 'name': CreateVM_Task, 'duration_secs': 0.532775} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.988137] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.988866] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.988957] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.989325] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.989589] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41fc2325-a964-4b27-9cbc-6028d4cafedf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.993833] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 808.993833] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522066e2-902e-cb0e-b650-893a372ec40e" [ 808.993833] env[61907]: _type = "Task" [ 808.993833] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.003502] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522066e2-902e-cb0e-b650-893a372ec40e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.037865] env[61907]: DEBUG oslo_vmware.api [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Task: {'id': task-1243770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179467} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.038215] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 809.038422] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 809.038607] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 809.038789] env[61907]: INFO nova.compute.manager [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Took 1.19 seconds to destroy the instance on the hypervisor. [ 809.039047] env[61907]: DEBUG oslo.service.loopingcall [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.039250] env[61907]: DEBUG nova.compute.manager [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 809.039367] env[61907]: DEBUG nova.network.neutron [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 809.041097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.041235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquired lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.041574] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.053729] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243771, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115465} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.053983] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.054938] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc86923-5313-48c2-8adb-fa7616af6639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.078806] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfiguring VM instance instance-00000039 to attach disk [datastore1] 52404147-0375-4f9c-9e1e-4a52001df7cb/52404147-0375-4f9c-9e1e-4a52001df7cb.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.081956] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d7ce539-a838-4f00-9356-4ec48e7e722d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.100734] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 809.100734] env[61907]: value = "task-1243772" [ 809.100734] env[61907]: _type = "Task" [ 809.100734] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.109474] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.120463] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a451c1cd-a339-4af5-b20d-f40380d3ff7b tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.605s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.171791] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da4b908-e50c-4919-ab0e-dbbaf242aaf4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.181082] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Suspending the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 809.181585] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-4502dc8a-02bc-493d-881e-f55de5d0a6b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.187516] env[61907]: DEBUG oslo_vmware.api [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 809.187516] env[61907]: value = "task-1243773" [ 809.187516] env[61907]: _type = "Task" [ 809.187516] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.196710] env[61907]: DEBUG oslo_vmware.api [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243773, 'name': SuspendVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.307208] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ffcc9d-ac8b-466d-965e-8e3b1af76d3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.311199] env[61907]: DEBUG oslo_concurrency.lockutils [req-469673d1-fa76-4fb0-a157-896ad86723b6 req-03e87ebd-7d2e-4900-b060-2cddacd946a9 service nova] Releasing lock "refresh_cache-f157c0db-054b-4d99-803c-11aa40cf0670" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.317278] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20777c2e-b989-4a01-af5e-7736b5319a94 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.358023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2fede0-aa8b-4d7b-9240-88b7dde6694e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.363362] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad25c03-0d71-4a68-b52d-840e64ed4429 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.379306] env[61907]: DEBUG nova.compute.provider_tree [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 809.508021] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522066e2-902e-cb0e-b650-893a372ec40e, 'name': SearchDatastore_Task, 'duration_secs': 0.010033} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.509165] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.509513] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.509801] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.509969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.510188] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.510761] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b114023-46a9-472d-934e-133a9e64096f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.520223] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.520506] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.521438] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33a8c6ee-e50a-4f7f-a04f-b13c62ec92ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.528037] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 809.528037] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52190da3-a780-1eeb-27d0-5e4ba893e1d2" [ 809.528037] env[61907]: _type = "Task" [ 809.528037] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.536863] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52190da3-a780-1eeb-27d0-5e4ba893e1d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.610392] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243772, 'name': ReconfigVM_Task, 'duration_secs': 0.461001} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.613785] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfigured VM instance instance-00000039 to attach disk [datastore1] 52404147-0375-4f9c-9e1e-4a52001df7cb/52404147-0375-4f9c-9e1e-4a52001df7cb.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 809.613785] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.613785] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-771ce51d-9cfe-40ed-920c-a9875157b9b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.619818] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 809.619818] env[61907]: value = "task-1243774" [ 809.619818] env[61907]: _type = "Task" [ 809.619818] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.622817] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 809.631379] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243774, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.639369] env[61907]: DEBUG nova.compute.manager [req-7172a6fb-425b-48d2-944c-78ebfd9dc836 req-3bf2661e-c56c-4f05-85b2-550ca59d71cc service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Received event network-vif-deleted-73705eb9-ea43-4aef-a8f6-68aa71f9582b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 809.639611] env[61907]: INFO nova.compute.manager [req-7172a6fb-425b-48d2-944c-78ebfd9dc836 req-3bf2661e-c56c-4f05-85b2-550ca59d71cc service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Neutron deleted interface 73705eb9-ea43-4aef-a8f6-68aa71f9582b; detaching it from the instance and deleting it from the info cache [ 809.639816] env[61907]: DEBUG nova.network.neutron [req-7172a6fb-425b-48d2-944c-78ebfd9dc836 req-3bf2661e-c56c-4f05-85b2-550ca59d71cc service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.705693] env[61907]: DEBUG oslo_vmware.api [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243773, 'name': SuspendVM_Task} progress is 58%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.858185] env[61907]: DEBUG nova.network.neutron [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.882158] env[61907]: DEBUG nova.scheduler.client.report [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 809.920092] env[61907]: DEBUG nova.network.neutron [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Updating instance_info_cache with network_info: [{"id": "30bd832d-f042-4736-8dea-152ff21d9739", "address": "fa:16:3e:fb:1e:56", "network": {"id": "f8a69d3c-5bdb-4061-b61c-8da44e9e3cdb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1045377162-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa11d39bdf14738a34f3534a44a9851", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bd832d-f0", "ovs_interfaceid": "30bd832d-f042-4736-8dea-152ff21d9739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.002865] env[61907]: DEBUG nova.network.neutron [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.039720] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52190da3-a780-1eeb-27d0-5e4ba893e1d2, 'name': SearchDatastore_Task, 'duration_secs': 0.011761} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.040679] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-750c7901-5e9e-4756-ac5e-1265549dcb6f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.046538] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 810.046538] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522aaed9-9b8d-9eca-7bda-f8287f7f593c" [ 810.046538] env[61907]: _type = "Task" [ 810.046538] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.055444] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522aaed9-9b8d-9eca-7bda-f8287f7f593c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.131710] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243774, 'name': Rename_Task, 'duration_secs': 0.159303} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.132091] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 810.134402] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-db3a2acf-5488-4d87-9c18-efcd307be83a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.141160] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 810.141160] env[61907]: value = "task-1243775" [ 810.141160] env[61907]: _type = "Task" [ 810.141160] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.144443] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-10f3595b-cfb1-4497-9ec0-49a146cd8d93 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.150845] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.152520] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.155515] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70aab678-b3ea-403a-bcce-e6fe46e02a41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.185832] env[61907]: DEBUG nova.compute.manager [req-7172a6fb-425b-48d2-944c-78ebfd9dc836 req-3bf2661e-c56c-4f05-85b2-550ca59d71cc service nova] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Detach interface failed, port_id=73705eb9-ea43-4aef-a8f6-68aa71f9582b, reason: Instance 5b20f363-dfad-4c2b-b757-4da4e5bd0b99 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 810.200823] env[61907]: DEBUG oslo_vmware.api [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243773, 'name': SuspendVM_Task, 'duration_secs': 0.749463} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.200823] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Suspended the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 810.200823] env[61907]: DEBUG nova.compute.manager [None req-c2c106ae-379f-4c93-b6d7-55a6142a6edb tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 810.203223] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2f031d-7566-4098-a87e-3866b4aa3281 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.361235] env[61907]: INFO nova.compute.manager [-] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Took 1.45 seconds to deallocate network for instance. [ 810.390687] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.390687] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 810.393018] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.878s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.395029] env[61907]: INFO nova.compute.claims [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.422970] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Releasing lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.424566] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Instance network_info: |[{"id": "30bd832d-f042-4736-8dea-152ff21d9739", "address": "fa:16:3e:fb:1e:56", "network": {"id": "f8a69d3c-5bdb-4061-b61c-8da44e9e3cdb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1045377162-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa11d39bdf14738a34f3534a44a9851", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bd832d-f0", "ovs_interfaceid": "30bd832d-f042-4736-8dea-152ff21d9739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 810.425701] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:1e:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '30bd832d-f042-4736-8dea-152ff21d9739', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 810.433958] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Creating folder: Project (bfa11d39bdf14738a34f3534a44a9851). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.434901] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3abbdaf-656b-46cf-a375-59bdf9970a8b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.447127] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Created folder: Project (bfa11d39bdf14738a34f3534a44a9851) in parent group-v268168. [ 810.447352] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Creating folder: Instances. Parent ref: group-v268224. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 810.447613] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b2adf48-7172-436b-93fd-a969067dc223 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.455652] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Created folder: Instances in parent group-v268224. [ 810.455913] env[61907]: DEBUG oslo.service.loopingcall [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.456153] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 810.456371] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-354a241c-1b48-4dda-9538-252a82c80c14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.476970] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 810.476970] env[61907]: value = "task-1243778" [ 810.476970] env[61907]: _type = "Task" [ 810.476970] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.484354] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243778, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.509506] env[61907]: INFO nova.compute.manager [-] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Took 1.47 seconds to deallocate network for instance. [ 810.558420] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522aaed9-9b8d-9eca-7bda-f8287f7f593c, 'name': SearchDatastore_Task, 'duration_secs': 0.014836} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.558796] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.559147] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f157c0db-054b-4d99-803c-11aa40cf0670/f157c0db-054b-4d99-803c-11aa40cf0670.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.559439] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2dc7aed-a37f-43e2-8a6d-32153f3d76bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.566982] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 810.566982] env[61907]: value = "task-1243779" [ 810.566982] env[61907]: _type = "Task" [ 810.566982] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.576213] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243779, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.615927] env[61907]: DEBUG nova.compute.manager [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Received event network-changed-30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 810.616179] env[61907]: DEBUG nova.compute.manager [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Refreshing instance network info cache due to event network-changed-30bd832d-f042-4736-8dea-152ff21d9739. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 810.616416] env[61907]: DEBUG oslo_concurrency.lockutils [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] Acquiring lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.616572] env[61907]: DEBUG oslo_concurrency.lockutils [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] Acquired lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.616759] env[61907]: DEBUG nova.network.neutron [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Refreshing network info cache for port 30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 810.651661] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243775, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.867536] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.903566] env[61907]: DEBUG nova.compute.utils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.905438] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 810.988291] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243778, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.016782] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.077418] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243779, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.47293} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.077857] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f157c0db-054b-4d99-803c-11aa40cf0670/f157c0db-054b-4d99-803c-11aa40cf0670.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 811.078212] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.078640] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a4f036b-18d9-4201-aecd-bc0e1524d3c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.088382] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 811.088382] env[61907]: value = "task-1243780" [ 811.088382] env[61907]: _type = "Task" [ 811.088382] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.094929] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243780, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.152909] env[61907]: DEBUG oslo_vmware.api [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243775, 'name': PowerOnVM_Task, 'duration_secs': 0.5115} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.153541] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 811.154469] env[61907]: INFO nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Took 8.42 seconds to spawn the instance on the hypervisor. [ 811.154469] env[61907]: DEBUG nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 811.155676] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad208a9-7024-4e18-9ad4-98a0dd80f0b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.406728] env[61907]: DEBUG nova.network.neutron [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Updated VIF entry in instance network info cache for port 30bd832d-f042-4736-8dea-152ff21d9739. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 811.407319] env[61907]: DEBUG nova.network.neutron [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Updating instance_info_cache with network_info: [{"id": "30bd832d-f042-4736-8dea-152ff21d9739", "address": "fa:16:3e:fb:1e:56", "network": {"id": "f8a69d3c-5bdb-4061-b61c-8da44e9e3cdb", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1045377162-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "bfa11d39bdf14738a34f3534a44a9851", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap30bd832d-f0", "ovs_interfaceid": "30bd832d-f042-4736-8dea-152ff21d9739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.408701] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 811.489100] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243778, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.597277] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243780, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065437} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.597850] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.598457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d89830-3d92-472e-8e10-cab4f82266a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.623335] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] f157c0db-054b-4d99-803c-11aa40cf0670/f157c0db-054b-4d99-803c-11aa40cf0670.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.625998] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20d78574-060a-40d1-9ada-fcf390c47f13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.664868] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 811.664868] env[61907]: value = "task-1243781" [ 811.664868] env[61907]: _type = "Task" [ 811.664868] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.680551] env[61907]: INFO nova.compute.manager [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Took 34.92 seconds to build instance. [ 811.688596] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243781, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.777232] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39e85cf-5940-4f34-b0d5-24db8c03a722 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.785671] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45beab12-923a-4f3a-9548-ed4402f7e0fd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.819727] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842f3191-b1b3-46d8-ae3c-cf57f66914c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.829093] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4c9758-d971-47b3-ac2b-4df5b3baa186 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.845053] env[61907]: DEBUG nova.compute.provider_tree [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.912972] env[61907]: DEBUG oslo_concurrency.lockutils [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] Releasing lock "refresh_cache-7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.912972] env[61907]: DEBUG nova.compute.manager [req-b9d56792-62de-4a09-9ea9-48c92bed389d req-5f94c6e1-ab41-4379-ab5a-2dfcf93ec070 service nova] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Received event network-vif-deleted-a486b310-da28-4ae5-929b-a4c6ab192b31 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 811.990805] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243778, 'name': CreateVM_Task, 'duration_secs': 1.480377} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.991169] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 811.992372] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.992838] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.993378] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 811.993789] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7631e1a1-24f7-4a61-a705-095bebebd7e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.002020] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 812.002020] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c08-24e8-9339-4a6a-24609ee78373" [ 812.002020] env[61907]: _type = "Task" [ 812.002020] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.009683] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c08-24e8-9339-4a6a-24609ee78373, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.178175] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243781, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.182783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cfa4fcbe-83e9-40c0-989a-0e1a73a1b587 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.165s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.347528] env[61907]: DEBUG nova.scheduler.client.report [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 812.419126] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 812.444264] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 812.444528] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 812.444695] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 812.444884] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 812.445118] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 812.445333] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 812.445599] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 812.445795] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 812.446031] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 812.446248] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 812.446480] env[61907]: DEBUG nova.virt.hardware [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 812.447397] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67e39dbf-8ef3-4351-947b-b560e6a10e9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.455948] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926c8a9f-733a-43d5-b1e2-5781af0fe2c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.474550] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.481502] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Creating folder: Project (606efe94f597469ab6f0a3f9738a416c). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.481916] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c2c6e8e2-f9ba-4469-8d3e-19c99fbf30c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.496455] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Created folder: Project (606efe94f597469ab6f0a3f9738a416c) in parent group-v268168. [ 812.496670] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Creating folder: Instances. Parent ref: group-v268227. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 812.496923] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52e30750-bc14-4ab2-8807-d6593e0ef6ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.513021] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c08-24e8-9339-4a6a-24609ee78373, 'name': SearchDatastore_Task, 'duration_secs': 0.010524} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.513021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.513213] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 812.513377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.513535] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.513765] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 812.514017] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Created folder: Instances in parent group-v268227. [ 812.514235] env[61907]: DEBUG oslo.service.loopingcall [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.514418] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-00964674-d309-4e3a-8d4f-4095ccb4ae38 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.516614] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 812.516827] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8273fab5-f416-4b59-ad62-32c6d029b7c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.537466] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 812.537466] env[61907]: value = "task-1243784" [ 812.537466] env[61907]: _type = "Task" [ 812.537466] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.542027] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 812.542027] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 812.543114] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cb0da22-0e97-42d7-ada9-864422db9859 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.550105] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243784, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.553838] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 812.553838] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52792a84-1081-b49e-8f4a-9e8da520b17a" [ 812.553838] env[61907]: _type = "Task" [ 812.553838] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.562729] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52792a84-1081-b49e-8f4a-9e8da520b17a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.656088] env[61907]: DEBUG nova.compute.manager [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-changed-ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 812.656430] env[61907]: DEBUG nova.compute.manager [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing instance network info cache due to event network-changed-ce4e1e2e-3671-4aba-a760-b96db6f21180. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 812.656642] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.656848] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.659143] env[61907]: DEBUG nova.network.neutron [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing network info cache for port ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 812.680530] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243781, 'name': ReconfigVM_Task, 'duration_secs': 0.952317} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.681098] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Reconfigured VM instance instance-0000003a to attach disk [datastore2] f157c0db-054b-4d99-803c-11aa40cf0670/f157c0db-054b-4d99-803c-11aa40cf0670.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.682566] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fe29f47c-e37e-486d-8837-6ff4a09fde5c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.687967] env[61907]: DEBUG nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 812.688384] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 812.692945] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e690d7f-d52e-4622-a4be-13cfa6a1f65d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.700014] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 812.700014] env[61907]: value = "task-1243785" [ 812.700014] env[61907]: _type = "Task" [ 812.700014] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.716497] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243785, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.854218] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.461s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.855258] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 812.857869] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.192s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.858163] env[61907]: DEBUG nova.objects.instance [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lazy-loading 'resources' on Instance uuid a6a117b3-7cc4-43c5-8202-12514f695e48 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 813.052934] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243784, 'name': CreateVM_Task, 'duration_secs': 0.323574} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.053327] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 813.053770] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.053945] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.054291] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 813.054602] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f7f8f29-f26d-436a-a1ba-90bc8500a0c1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.067023] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52792a84-1081-b49e-8f4a-9e8da520b17a, 'name': SearchDatastore_Task, 'duration_secs': 0.010802} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.068023] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 813.068023] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ae3bf2-cd2e-265a-9366-0686ba46e4fd" [ 813.068023] env[61907]: _type = "Task" [ 813.068023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.068312] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-974cb29b-5298-471d-93d5-6bb5da34d02e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.078549] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 813.078549] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c499a7-0f49-3eaa-e6c6-a8841375a84a" [ 813.078549] env[61907]: _type = "Task" [ 813.078549] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.082851] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ae3bf2-cd2e-265a-9366-0686ba46e4fd, 'name': SearchDatastore_Task, 'duration_secs': 0.010984} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.085934] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.086159] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 813.086507] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.093127] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c499a7-0f49-3eaa-e6c6-a8841375a84a, 'name': SearchDatastore_Task, 'duration_secs': 0.010077} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.093386] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.093640] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111/7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.093931] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.094134] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 813.094357] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0febefee-7e51-484e-83bf-7e90f0d5fc7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.096299] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d48d08de-22ac-4762-9912-6dba3e2dc776 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.106712] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 813.106712] env[61907]: value = "task-1243786" [ 813.106712] env[61907]: _type = "Task" [ 813.106712] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.106903] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 813.107066] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 813.110758] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bb08942-5b23-4c12-ac51-2c03d93e727f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.118491] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 813.118491] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5247eff3-ff69-592b-6838-4eb15ecb7461" [ 813.118491] env[61907]: _type = "Task" [ 813.118491] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.121300] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243786, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.130215] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5247eff3-ff69-592b-6838-4eb15ecb7461, 'name': SearchDatastore_Task, 'duration_secs': 0.010508} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.130969] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28d3f376-779d-4bf1-b53c-1c9bbb1a79a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.138120] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 813.138120] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5294196c-a2bc-2240-7bb9-5c2c24a335a6" [ 813.138120] env[61907]: _type = "Task" [ 813.138120] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.147453] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5294196c-a2bc-2240-7bb9-5c2c24a335a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.211139] env[61907]: INFO nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] instance snapshotting [ 813.211491] env[61907]: WARNING nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 813.214900] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe3f04c-da63-4572-a2cb-07ee25e7bfe0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.221317] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243785, 'name': Rename_Task, 'duration_secs': 0.163378} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.221651] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 813.221974] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ddac2d44-69e6-4edd-bbdc-38c3de89ec52 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.238910] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.240414] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-384c0c4c-d11e-4e21-848e-9235fd8c4c58 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.246726] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 813.246726] env[61907]: value = "task-1243787" [ 813.246726] env[61907]: _type = "Task" [ 813.246726] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.268818] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243787, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.363543] env[61907]: DEBUG nova.compute.utils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.365943] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 813.365943] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 813.444871] env[61907]: DEBUG nova.policy [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0745691ba6c44a42b887c649d7752ebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2c808b1baf1842fbb9e2d28f0031e4d7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.555129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.555129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.555129] env[61907]: INFO nova.compute.manager [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Shelving [ 813.590552] env[61907]: DEBUG nova.network.neutron [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updated VIF entry in instance network info cache for port ce4e1e2e-3671-4aba-a760-b96db6f21180. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 813.590882] env[61907]: DEBUG nova.network.neutron [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.620181] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243786, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49814} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.620485] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111/7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 813.620710] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.620973] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61c309fd-daba-494f-86cf-66013439f978 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.627738] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 813.627738] env[61907]: value = "task-1243788" [ 813.627738] env[61907]: _type = "Task" [ 813.627738] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.639757] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243788, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.671227] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5294196c-a2bc-2240-7bb9-5c2c24a335a6, 'name': SearchDatastore_Task, 'duration_secs': 0.009678} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.671562] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.671800] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 813.672087] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-964361e5-0656-4e4e-b201-d9c66e3298f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.680676] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 813.680676] env[61907]: value = "task-1243789" [ 813.680676] env[61907]: _type = "Task" [ 813.680676] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.692735] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243789, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.745619] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3d39e4-8a25-4ffe-8519-3583fae19a42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.757023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9dcef49-3f52-4c4b-9fa8-7bf17c070767 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.764164] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 813.764649] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d0e5ec30-6616-4a70-8a70-ea4a82c41f4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.771081] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243787, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.798661] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Successfully created port: 718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 813.801194] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f634b0-7fb5-49af-b085-c40bcb902f17 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.806676] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 813.806676] env[61907]: value = "task-1243790" [ 813.806676] env[61907]: _type = "Task" [ 813.806676] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.814236] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b563edad-8871-44d3-9761-0b3f5c05711e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.821149] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243790, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.832182] env[61907]: DEBUG nova.compute.provider_tree [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.868981] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 814.096081] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0b7b51-7e15-42f4-aba2-490b0f875731 req-4bd5a27d-4325-4b0b-9be4-0ce82ca55203 service nova] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.141859] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243788, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085368} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.142270] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 814.143191] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4486704e-6de4-4472-8a61-5f42e07a44ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.179028] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111/7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 814.179524] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-72739356-7b81-4cad-ad6a-93af22d330a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.212025] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243789, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.212025] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 814.212025] env[61907]: value = "task-1243791" [ 814.212025] env[61907]: _type = "Task" [ 814.212025] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.224025] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243791, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.263649] env[61907]: DEBUG oslo_vmware.api [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243787, 'name': PowerOnVM_Task, 'duration_secs': 0.749881} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.264208] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 814.264599] env[61907]: INFO nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Took 8.95 seconds to spawn the instance on the hypervisor. [ 814.264963] env[61907]: DEBUG nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 814.266229] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb71e36-0124-408f-b683-2580ab20107c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.317235] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243790, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.336197] env[61907]: DEBUG nova.scheduler.client.report [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 814.566684] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 814.566800] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c915247-6218-456e-95cf-c71e41191dfb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.576760] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 814.576760] env[61907]: value = "task-1243792" [ 814.576760] env[61907]: _type = "Task" [ 814.576760] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.586678] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243792, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.706338] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243789, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587851} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.706712] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 814.706950] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 814.707278] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97a30118-d5ef-4e01-bca4-5aa86ca7e1c1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.716168] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 814.716168] env[61907]: value = "task-1243793" [ 814.716168] env[61907]: _type = "Task" [ 814.716168] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.719624] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243791, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.729541] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243793, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.785566] env[61907]: INFO nova.compute.manager [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Took 35.19 seconds to build instance. [ 814.819215] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243790, 'name': CreateSnapshot_Task, 'duration_secs': 0.692369} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.819507] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 814.820321] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef739f8c-b89c-4303-9e12-f6917ef44c3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.841048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.983s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.843714] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.655s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.847513] env[61907]: INFO nova.compute.claims [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 814.865832] env[61907]: INFO nova.scheduler.client.report [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted allocations for instance a6a117b3-7cc4-43c5-8202-12514f695e48 [ 814.881089] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 814.908659] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.909011] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.909227] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.909445] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.909657] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.909825] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.910267] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.910267] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.910405] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.910482] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.910664] env[61907]: DEBUG nova.virt.hardware [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.911621] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9190bb7f-6a26-42ea-b8e1-080e6cb04b06 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.921123] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168bc55d-f6aa-49f0-83fe-3846759b764a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.087650] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243792, 'name': PowerOffVM_Task, 'duration_secs': 0.277764} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.087873] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.088684] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d620c0b-b3d1-4bc9-9238-ab620a6b4203 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.108629] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc34064-c981-40f4-9b57-2f6155114d7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.221388] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243791, 'name': ReconfigVM_Task, 'duration_secs': 0.823156} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.224568] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111/7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.225260] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d81d71e4-2f70-4a85-9009-b706c0c97146 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.232217] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243793, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074923} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.232569] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 815.235064] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c504665a-de83-4b5f-baf9-8e456ab26018 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.237639] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 815.237639] env[61907]: value = "task-1243794" [ 815.237639] env[61907]: _type = "Task" [ 815.237639] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.259302] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 815.260748] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a73b853c-cebe-4e95-9f10-6795eba93392 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.279069] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243794, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.286281] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 815.286281] env[61907]: value = "task-1243795" [ 815.286281] env[61907]: _type = "Task" [ 815.286281] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.289816] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5c8e469d-f71a-42a4-9ffc-c912ce7bd57e tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.154s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.296869] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243795, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.340679] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 815.341418] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-550d9185-5d0c-4db6-b668-b1f54545a270 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.351648] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 815.351648] env[61907]: value = "task-1243796" [ 815.351648] env[61907]: _type = "Task" [ 815.351648] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.362770] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243796, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.373911] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf4c99b0-db33-418b-95fc-6330b4590af2 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "a6a117b3-7cc4-43c5-8202-12514f695e48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.106s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.418449] env[61907]: DEBUG nova.compute.manager [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received event network-vif-plugged-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 815.418777] env[61907]: DEBUG oslo_concurrency.lockutils [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] Acquiring lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.419489] env[61907]: DEBUG oslo_concurrency.lockutils [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] Lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.419489] env[61907]: DEBUG oslo_concurrency.lockutils [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] Lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.419773] env[61907]: DEBUG nova.compute.manager [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] No waiting events found dispatching network-vif-plugged-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 815.420413] env[61907]: WARNING nova.compute.manager [req-960bdb21-dbd4-4766-9b9d-808a2bf1ba56 req-9e530160-071a-4297-98ca-bd3b2782a96d service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received unexpected event network-vif-plugged-718419f3-2250-41d9-88ef-464d9384887a for instance with vm_state building and task_state spawning. [ 815.499788] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "f157c0db-054b-4d99-803c-11aa40cf0670" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.500116] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.500361] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.500563] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.500783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.503375] env[61907]: INFO nova.compute.manager [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Terminating instance [ 815.536170] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Successfully updated port: 718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 815.620431] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 815.620757] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5ee2d26d-2bb4-4014-8dbc-c9e3ce3e2019 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.628496] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 815.628496] env[61907]: value = "task-1243797" [ 815.628496] env[61907]: _type = "Task" [ 815.628496] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.637654] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243797, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.748567] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243794, 'name': Rename_Task, 'duration_secs': 0.247059} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.749060] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 815.749968] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bd9d17f-8116-4fd2-bd6b-1bdd88325c6a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.758023] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 815.758023] env[61907]: value = "task-1243798" [ 815.758023] env[61907]: _type = "Task" [ 815.758023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.764720] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243798, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.792983] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 815.799290] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243795, 'name': ReconfigVM_Task, 'duration_secs': 0.421738} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.799885] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Reconfigured VM instance instance-0000003c to attach disk [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 815.800598] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-659cff36-63d1-44f3-8437-6cce1aa04941 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.816278] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 815.816278] env[61907]: value = "task-1243799" [ 815.816278] env[61907]: _type = "Task" [ 815.816278] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.826486] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243799, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.863788] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243796, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.011395] env[61907]: DEBUG nova.compute.manager [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 816.011931] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 816.016016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ef033-21dd-4805-9b30-06a5aa0e55c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.033771] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 816.033771] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f66e9fd-8261-41d2-b853-60207feceee9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.041827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.041827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.041827] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.042851] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 816.042851] env[61907]: value = "task-1243800" [ 816.042851] env[61907]: _type = "Task" [ 816.042851] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.054581] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.140484] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243797, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.267055] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243798, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.272523] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f121827-475e-416f-8c88-1fb0e4839dff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.282048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43ce2cf-1819-49f6-986d-ae3dc271f409 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.335412] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25f4f96-c1fe-4bae-8a12-95432776b7b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.349059] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a106e74-01ff-4dbb-b258-54c5e4946316 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.353457] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243799, 'name': Rename_Task, 'duration_secs': 0.154938} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.354480] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.354955] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 816.358510] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1dac5a6-1a14-4eb1-b12c-58c4b4a34e1a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.369175] env[61907]: DEBUG nova.compute.provider_tree [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.376571] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 816.376571] env[61907]: value = "task-1243801" [ 816.376571] env[61907]: _type = "Task" [ 816.376571] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.376571] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243796, 'name': CloneVM_Task} progress is 95%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.387135] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243801, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.554812] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243800, 'name': PowerOffVM_Task, 'duration_secs': 0.160612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.555227] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 816.555322] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 816.555528] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-48c7f4a8-5cf7-4f24-9ede-96c338209cd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.578774] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 816.639633] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243797, 'name': CreateSnapshot_Task, 'duration_secs': 0.753191} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.640046] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 816.640873] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d55177-04d0-4be4-8512-6862328b3235 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.769844] env[61907]: DEBUG oslo_vmware.api [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243798, 'name': PowerOnVM_Task, 'duration_secs': 0.596832} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.770193] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 816.770406] env[61907]: INFO nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Took 8.93 seconds to spawn the instance on the hypervisor. [ 816.770592] env[61907]: DEBUG nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 816.772281] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718b2719-db10-41be-b65a-db06d5af1454 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.775950] env[61907]: DEBUG nova.network.neutron [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.815751] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 816.816150] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 816.816470] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Deleting the datastore file [datastore2] f157c0db-054b-4d99-803c-11aa40cf0670 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 816.817828] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bcd526c2-4554-40bb-86e4-7ff6fcd51800 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.824848] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for the task: (returnval){ [ 816.824848] env[61907]: value = "task-1243803" [ 816.824848] env[61907]: _type = "Task" [ 816.824848] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.833695] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.876237] env[61907]: DEBUG nova.scheduler.client.report [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 816.880107] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243796, 'name': CloneVM_Task, 'duration_secs': 1.320742} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.883994] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Created linked-clone VM from snapshot [ 816.884819] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3101104-1856-48f0-bccd-5ddef1a52a22 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.894776] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243801, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.901450] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Uploading image 39c1a265-6dd8-4c56-ac69-1ca1342620d3 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 816.927954] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 816.927954] env[61907]: value = "vm-268231" [ 816.927954] env[61907]: _type = "VirtualMachine" [ 816.927954] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 816.928380] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-819bb9e4-589a-4aef-bc8a-98978f0e9470 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.935150] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease: (returnval){ [ 816.935150] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52647c64-92d8-2879-b4ea-62b1e1d5f069" [ 816.935150] env[61907]: _type = "HttpNfcLease" [ 816.935150] env[61907]: } obtained for exporting VM: (result){ [ 816.935150] env[61907]: value = "vm-268231" [ 816.935150] env[61907]: _type = "VirtualMachine" [ 816.935150] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 816.935420] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the lease: (returnval){ [ 816.935420] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52647c64-92d8-2879-b4ea-62b1e1d5f069" [ 816.935420] env[61907]: _type = "HttpNfcLease" [ 816.935420] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 816.942437] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 816.942437] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52647c64-92d8-2879-b4ea-62b1e1d5f069" [ 816.942437] env[61907]: _type = "HttpNfcLease" [ 816.942437] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 817.158644] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 817.159029] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-9397cda2-fb38-4985-afb9-1a11a302f242 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.168679] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 817.168679] env[61907]: value = "task-1243805" [ 817.168679] env[61907]: _type = "Task" [ 817.168679] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.177662] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243805, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.224613] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.224852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.278164] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.278508] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Instance network_info: |[{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 817.278951] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:10:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '62d6a386-ffdb-4232-83f3-cb21c5e59e85', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '718419f3-2250-41d9-88ef-464d9384887a', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.290235] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Creating folder: Project (2c808b1baf1842fbb9e2d28f0031e4d7). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.295823] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d41fa448-a4b7-40e6-be12-a3d5be86615a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.301285] env[61907]: INFO nova.compute.manager [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Took 36.16 seconds to build instance. [ 817.307661] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Created folder: Project (2c808b1baf1842fbb9e2d28f0031e4d7) in parent group-v268168. [ 817.307909] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Creating folder: Instances. Parent ref: group-v268234. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 817.308229] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d80ca369-d617-4b92-8cd5-e919718b60dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.319037] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Created folder: Instances in parent group-v268234. [ 817.319348] env[61907]: DEBUG oslo.service.loopingcall [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.319883] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.320173] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e470fd6-2d0d-47dd-87ce-8fc3e143aa20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.344138] env[61907]: DEBUG oslo_vmware.api [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Task: {'id': task-1243803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162456} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.345507] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 817.345714] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 817.345895] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 817.346088] env[61907]: INFO nova.compute.manager [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Took 1.33 seconds to destroy the instance on the hypervisor. [ 817.346341] env[61907]: DEBUG oslo.service.loopingcall [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.346535] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.346535] env[61907]: value = "task-1243808" [ 817.346535] env[61907]: _type = "Task" [ 817.346535] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.346730] env[61907]: DEBUG nova.compute.manager [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 817.346846] env[61907]: DEBUG nova.network.neutron [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.357372] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243808, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.381828] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.382387] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 817.390807] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 24.844s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.391068] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.391279] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 817.391633] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.114s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.393612] env[61907]: INFO nova.compute.claims [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.396923] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22259083-db84-4ef6-9db1-1887ef23a6d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.405568] env[61907]: DEBUG oslo_vmware.api [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243801, 'name': PowerOnVM_Task, 'duration_secs': 0.826105} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.407327] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 817.407622] env[61907]: INFO nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Took 4.99 seconds to spawn the instance on the hypervisor. [ 817.407830] env[61907]: DEBUG nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 817.409505] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554bacb9-adc3-48c8-9db9-1cd6928812e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.412818] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c029256-70e2-47d4-b0bb-3041a556c095 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.433728] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f80562-e624-404b-a999-d759ca1be729 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.444935] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 817.444935] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52647c64-92d8-2879-b4ea-62b1e1d5f069" [ 817.444935] env[61907]: _type = "HttpNfcLease" [ 817.444935] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 817.445557] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 817.445557] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52647c64-92d8-2879-b4ea-62b1e1d5f069" [ 817.445557] env[61907]: _type = "HttpNfcLease" [ 817.445557] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 817.446519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f72df7-0c6e-482f-a96a-6ab180a69342 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.450009] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c879ee43-d565-47c8-8e1a-e02a17b53826 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.458696] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 817.458696] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 817.487557] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181219MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 817.487740] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.582352] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-67047124-e363-486a-9865-954a12161047 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.648599] env[61907]: DEBUG nova.compute.manager [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received event network-changed-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 817.648813] env[61907]: DEBUG nova.compute.manager [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing instance network info cache due to event network-changed-718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 817.649064] env[61907]: DEBUG oslo_concurrency.lockutils [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] Acquiring lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.649218] env[61907]: DEBUG oslo_concurrency.lockutils [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] Acquired lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.649422] env[61907]: DEBUG nova.network.neutron [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing network info cache for port 718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 817.679923] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243805, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.803127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7f6ecc79-5402-453f-828d-e3ab3c5e45dd tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 93.941s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.860377] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243808, 'name': CreateVM_Task, 'duration_secs': 0.416878} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.860688] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.861700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.861700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.862516] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.863186] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-156212c6-61ed-4dc1-af68-784f519e1ebd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.867951] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 817.867951] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52710839-c99a-f91a-9ac5-36fc7caf2e01" [ 817.867951] env[61907]: _type = "Task" [ 817.867951] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.876918] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52710839-c99a-f91a-9ac5-36fc7caf2e01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.890169] env[61907]: DEBUG nova.compute.utils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.891822] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 817.892100] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.944926] env[61907]: INFO nova.compute.manager [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Took 32.62 seconds to build instance. [ 817.957693] env[61907]: DEBUG nova.policy [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26bbcb77d3874106808300b1bdb36a8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c53d877a81a7422db4fdfbbc6ed37444', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 818.181996] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243805, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.304754] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Successfully created port: 0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 818.308020] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 818.379948] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52710839-c99a-f91a-9ac5-36fc7caf2e01, 'name': SearchDatastore_Task, 'duration_secs': 0.009381} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.380363] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.380731] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.381099] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.381466] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.381841] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.385104] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f550553-e4c0-48ce-a17f-6f4fc80fb2c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.394864] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 818.398405] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.398951] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.409184] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9282df5-e3d9-41c2-ac9f-42f59c90caeb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.419671] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 818.419671] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5284e66c-cd6f-dbdd-e636-f59e362b26be" [ 818.419671] env[61907]: _type = "Task" [ 818.419671] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.436072] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5284e66c-cd6f-dbdd-e636-f59e362b26be, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.439270] env[61907]: DEBUG nova.network.neutron [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updated VIF entry in instance network info cache for port 718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 818.439766] env[61907]: DEBUG nova.network.neutron [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.441250] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77d09ac-6e34-44af-bbd5-14ecbcf6cf19 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.449284] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9d7bea8-653b-419c-8645-46fbfc9f5c9f tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.546s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.452237] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 818.452237] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a4702d-18a8-4b8b-0593-530ccf1831f9" [ 818.452237] env[61907]: _type = "Task" [ 818.452237] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.463995] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a4702d-18a8-4b8b-0593-530ccf1831f9, 'name': SearchDatastore_Task, 'duration_secs': 0.01002} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.469152] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.469492] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5c908f9c-7efb-4588-b6bf-70babb173767/5c908f9c-7efb-4588-b6bf-70babb173767.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 818.470293] env[61907]: DEBUG nova.network.neutron [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.471942] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7251a8a-0569-48c7-b40c-a42706710ee2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.481291] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 818.481291] env[61907]: value = "task-1243809" [ 818.481291] env[61907]: _type = "Task" [ 818.481291] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.492045] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.703353] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243805, 'name': CloneVM_Task, 'duration_secs': 1.471386} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.707624] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Created linked-clone VM from snapshot [ 818.711377] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f491037-6a9b-40d6-890c-355806f1a860 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.724881] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Uploading image c525c73f-cabe-4077-be2c-3ee109d7c7c8 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 818.762062] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 818.762062] env[61907]: value = "vm-268233" [ 818.762062] env[61907]: _type = "VirtualMachine" [ 818.762062] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 818.762917] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-77fcc476-d430-4efe-a54b-2b98cbe6f437 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.771800] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lease: (returnval){ [ 818.771800] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52311d50-b8d9-c18f-43bb-ef108daca26e" [ 818.771800] env[61907]: _type = "HttpNfcLease" [ 818.771800] env[61907]: } obtained for exporting VM: (result){ [ 818.771800] env[61907]: value = "vm-268233" [ 818.771800] env[61907]: _type = "VirtualMachine" [ 818.771800] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 818.772537] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the lease: (returnval){ [ 818.772537] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52311d50-b8d9-c18f-43bb-ef108daca26e" [ 818.772537] env[61907]: _type = "HttpNfcLease" [ 818.772537] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 818.782543] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 818.782543] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52311d50-b8d9-c18f-43bb-ef108daca26e" [ 818.782543] env[61907]: _type = "HttpNfcLease" [ 818.782543] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 818.838612] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.951689] env[61907]: DEBUG oslo_concurrency.lockutils [req-275737dc-1545-4c15-9a2d-84e74c38a88b req-4609f6d1-d933-464f-99d5-8154dd920497 service nova] Releasing lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.958445] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 818.973256] env[61907]: INFO nova.compute.manager [-] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Took 1.63 seconds to deallocate network for instance. [ 818.994099] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243809, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.002302] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb40d81-a9c5-4205-88f9-318aaad49b18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.010085] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc032c5-ca9c-4a6f-8000-fe1cd58a0386 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.044570] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bacc6b9-c17f-4ddb-a8f6-31d78e76c866 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.053503] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4659e103-fdb8-46b2-a381-19a255e73cbd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.068485] env[61907]: DEBUG nova.compute.provider_tree [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.203256] env[61907]: INFO nova.compute.manager [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Rebuilding instance [ 819.246594] env[61907]: DEBUG nova.compute.manager [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 819.247675] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2724bbc-4042-4871-8810-88ac3f4d2357 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.282694] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 819.282694] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52311d50-b8d9-c18f-43bb-ef108daca26e" [ 819.282694] env[61907]: _type = "HttpNfcLease" [ 819.282694] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 819.282694] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 819.282694] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52311d50-b8d9-c18f-43bb-ef108daca26e" [ 819.282694] env[61907]: _type = "HttpNfcLease" [ 819.282694] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 819.283081] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a243d5e6-f04e-417b-b0bb-b6a70532ff56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.291396] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 819.291629] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 819.393108] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c69c17b6-0f2f-43e8-8456-cc824198584c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.413478] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 819.481422] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.482716] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.493258] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243809, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.571343] env[61907]: DEBUG nova.scheduler.client.report [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 819.599935] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 819.600321] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 819.600556] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 819.600844] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 819.601085] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 819.601329] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 819.601673] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 819.601869] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 819.602108] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 819.602325] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 819.602560] env[61907]: DEBUG nova.virt.hardware [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 819.603515] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f3deb5-7018-4423-ac29-23c5e3fa3bd8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.615946] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64240236-a1f0-4957-81cc-e7535378ce9a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.682838] env[61907]: DEBUG nova.compute.manager [req-e8cf0be3-524c-4121-84a3-1f557364ca17 req-c3f0a322-05dd-4e3e-8d24-848298548a17 service nova] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Received event network-vif-deleted-369b29da-86e9-48a7-8cae-7627cd00ce30 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 819.865324] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.868376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.868376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.868376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.868376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.869785] env[61907]: INFO nova.compute.manager [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Terminating instance [ 819.907791] env[61907]: DEBUG nova.compute.manager [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Received event network-vif-plugged-0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 819.908583] env[61907]: DEBUG oslo_concurrency.lockutils [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] Acquiring lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.909051] env[61907]: DEBUG oslo_concurrency.lockutils [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.909435] env[61907]: DEBUG oslo_concurrency.lockutils [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.909802] env[61907]: DEBUG nova.compute.manager [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] No waiting events found dispatching network-vif-plugged-0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 819.910190] env[61907]: WARNING nova.compute.manager [req-dd74eec3-a2b1-4838-99b1-ef19ff541889 req-9664151b-c7b5-4a38-b308-4f33bb5a2681 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Received unexpected event network-vif-plugged-0f0768ac-bf89-4dc8-ba37-3e163bd551b1 for instance with vm_state building and task_state spawning. [ 819.993420] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243809, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.503693} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.995175] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5c908f9c-7efb-4588-b6bf-70babb173767/5c908f9c-7efb-4588-b6bf-70babb173767.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.995175] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.995470] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f31cd85-9c2b-4783-94ef-4cd42e964712 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.004169] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 820.004169] env[61907]: value = "task-1243811" [ 820.004169] env[61907]: _type = "Task" [ 820.004169] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.014424] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.020347] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Successfully updated port: 0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 820.077551] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.686s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.078227] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 820.081353] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.325s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.081878] env[61907]: DEBUG nova.objects.instance [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 820.261409] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.262165] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70603a9a-ad3a-4636-96a0-f5e0ef1d38ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.270253] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 820.270253] env[61907]: value = "task-1243812" [ 820.270253] env[61907]: _type = "Task" [ 820.270253] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.280645] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.374631] env[61907]: DEBUG nova.compute.manager [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 820.374867] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.375789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ecce1d1-9b10-4484-816e-34bba5c09c96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.383789] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 820.384273] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8556b66a-1a21-4f85-8ef7-b7f0f93a35dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.391797] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 820.391797] env[61907]: value = "task-1243813" [ 820.391797] env[61907]: _type = "Task" [ 820.391797] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.399948] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243813, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.515309] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078107} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.515700] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.516666] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb2c180-a6b5-4621-952c-06649b557088 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.532377] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.532787] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.533458] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.543172] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Reconfiguring VM instance instance-0000003d to attach disk [datastore2] 5c908f9c-7efb-4588-b6bf-70babb173767/5c908f9c-7efb-4588-b6bf-70babb173767.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.546647] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84c5393d-adb1-4474-8e9c-5e7f8cd2b629 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.568952] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 820.568952] env[61907]: value = "task-1243814" [ 820.568952] env[61907]: _type = "Task" [ 820.568952] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.578536] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243814, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.587904] env[61907]: DEBUG nova.compute.utils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.592728] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 820.592957] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.595670] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.646041] env[61907]: DEBUG nova.policy [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88b29a917e03445281f16fba6d536280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7652e98cde994af28b7bac0b81547474', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.787802] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243812, 'name': PowerOffVM_Task, 'duration_secs': 0.206614} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.788435] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.788928] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.789860] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11ef292-4a0a-4dce-9390-ba923edeac4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.797876] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.798169] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-195c6647-42a3-4671-afd1-18db10302be8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.800628] env[61907]: DEBUG nova.network.neutron [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Updating instance_info_cache with network_info: [{"id": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "address": "fa:16:3e:15:81:45", "network": {"id": "72104f6a-7f92-428f-9f2f-ed6a9a42fbd9", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-14201096-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c53d877a81a7422db4fdfbbc6ed37444", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f0768ac-bf", "ovs_interfaceid": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.829654] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.831046] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.831046] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Deleting the datastore file [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.831046] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3ba42fd-ecea-4e02-bf8b-5a884d9d1cd0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.837960] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 820.837960] env[61907]: value = "task-1243816" [ 820.837960] env[61907]: _type = "Task" [ 820.837960] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.847256] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243816, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.902544] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243813, 'name': PowerOffVM_Task, 'duration_secs': 0.247267} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.903047] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 820.903327] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 820.903664] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e7b0866-8f44-46f8-b016-054b304fb1b7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.972115] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 820.972115] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 820.972115] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Deleting the datastore file [datastore2] 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 820.972115] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fcbfaa8-6f87-46db-8b23-9a92b2a84341 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.979063] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for the task: (returnval){ [ 820.979063] env[61907]: value = "task-1243818" [ 820.979063] env[61907]: _type = "Task" [ 820.979063] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.987914] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.988773] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Successfully created port: f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.078898] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243814, 'name': ReconfigVM_Task, 'duration_secs': 0.430344} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.079405] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Reconfigured VM instance instance-0000003d to attach disk [datastore2] 5c908f9c-7efb-4588-b6bf-70babb173767/5c908f9c-7efb-4588-b6bf-70babb173767.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.080175] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e949a0a-cbc4-4f5c-943e-6fe048e1a34e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.087472] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 821.087472] env[61907]: value = "task-1243819" [ 821.087472] env[61907]: _type = "Task" [ 821.087472] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.093043] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 821.101181] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1a97a854-dd29-4481-82ff-ef87374e39db tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.020s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.103090] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243819, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.103824] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.143s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.104190] env[61907]: DEBUG nova.objects.instance [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'resources' on Instance uuid 1ef8594c-5d9a-491f-89f6-e5d7f35ce751 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 821.303687] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Releasing lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.304207] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Instance network_info: |[{"id": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "address": "fa:16:3e:15:81:45", "network": {"id": "72104f6a-7f92-428f-9f2f-ed6a9a42fbd9", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-14201096-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c53d877a81a7422db4fdfbbc6ed37444", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f0768ac-bf", "ovs_interfaceid": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 821.304975] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:81:45', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd47d5e1d-e66d-4f2c-83e6-d5e78c2b767d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0f0768ac-bf89-4dc8-ba37-3e163bd551b1', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 821.314211] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Creating folder: Project (c53d877a81a7422db4fdfbbc6ed37444). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.314835] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86d55d34-2556-4dcd-8f16-c02a240714b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.325243] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Created folder: Project (c53d877a81a7422db4fdfbbc6ed37444) in parent group-v268168. [ 821.325494] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Creating folder: Instances. Parent ref: group-v268237. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 821.325856] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-546878cd-2b2d-42f1-9cde-0b7ece29778c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.335657] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Created folder: Instances in parent group-v268237. [ 821.336056] env[61907]: DEBUG oslo.service.loopingcall [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.336196] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 821.336771] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ad5dbffb-ef23-492b-affa-2eb93ed0641d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.359994] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.359994] env[61907]: value = "task-1243822" [ 821.359994] env[61907]: _type = "Task" [ 821.359994] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.363011] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243816, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139396} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.366402] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.366677] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.366926] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.375664] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243822, 'name': CreateVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.491027] env[61907]: DEBUG oslo_vmware.api [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Task: {'id': task-1243818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.307295} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.491396] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 821.491683] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 821.491889] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 821.492191] env[61907]: INFO nova.compute.manager [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Took 1.12 seconds to destroy the instance on the hypervisor. [ 821.492493] env[61907]: DEBUG oslo.service.loopingcall [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 821.492749] env[61907]: DEBUG nova.compute.manager [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 821.492864] env[61907]: DEBUG nova.network.neutron [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.598332] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243819, 'name': Rename_Task, 'duration_secs': 0.215221} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.598641] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.599075] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5ea04414-a404-4942-be0e-4942697b6783 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.615605] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 821.615605] env[61907]: value = "task-1243823" [ 821.615605] env[61907]: _type = "Task" [ 821.615605] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.633245] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243823, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.880986] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243822, 'name': CreateVM_Task, 'duration_secs': 0.318794} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.881180] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 821.882072] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.882262] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.882803] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.883478] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fa789f1-0dff-483c-93be-695e7d245d74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.894582] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 821.894582] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dff04e-60eb-7cfd-b630-9d17d764c8ad" [ 821.894582] env[61907]: _type = "Task" [ 821.894582] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.907353] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dff04e-60eb-7cfd-b630-9d17d764c8ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.976781] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97623e21-4338-465a-85e3-e09d274ce240 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.985221] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853ff8ce-6e08-459f-8109-a8f2782b35ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.020651] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77116afd-4013-4897-9753-c46d6306acee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.032140] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-862e366f-9b59-4f13-bed7-dd619cf89f64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.046805] env[61907]: DEBUG nova.compute.provider_tree [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.113050] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 822.126585] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243823, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.257987] env[61907]: DEBUG nova.compute.manager [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Received event network-changed-0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 822.258517] env[61907]: DEBUG nova.compute.manager [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Refreshing instance network info cache due to event network-changed-0f0768ac-bf89-4dc8-ba37-3e163bd551b1. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 822.258616] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] Acquiring lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.258775] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] Acquired lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.258947] env[61907]: DEBUG nova.network.neutron [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Refreshing network info cache for port 0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 822.392459] env[61907]: DEBUG nova.network.neutron [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.409383] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dff04e-60eb-7cfd-b630-9d17d764c8ad, 'name': SearchDatastore_Task, 'duration_secs': 0.009851} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.409794] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.410925] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.410925] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.410925] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.410925] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.411691] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04c23e2d-42bf-43e8-bb43-1ab1856b6ef7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.420139] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.420389] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 822.421230] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb4fed63-9b80-45dd-8eb9-5dcad8b5a704 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.427702] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 822.427702] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5228ddaa-0250-cae1-08e5-bcb4205a5e25" [ 822.427702] env[61907]: _type = "Task" [ 822.427702] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.436329] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5228ddaa-0250-cae1-08e5-bcb4205a5e25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.551183] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Successfully updated port: f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 822.553596] env[61907]: DEBUG nova.scheduler.client.report [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 822.627926] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243823, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.659173] env[61907]: DEBUG nova.compute.manager [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Received event network-vif-plugged-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 822.659255] env[61907]: DEBUG oslo_concurrency.lockutils [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.659504] env[61907]: DEBUG oslo_concurrency.lockutils [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.660267] env[61907]: DEBUG oslo_concurrency.lockutils [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.660555] env[61907]: DEBUG nova.compute.manager [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] No waiting events found dispatching network-vif-plugged-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 822.660798] env[61907]: WARNING nova.compute.manager [req-d505d139-6464-47ec-80e4-99717d0d6a2b req-d052d4f9-afb6-4eb7-ae83-cd83f60a5e0d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Received unexpected event network-vif-plugged-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 for instance with vm_state building and task_state spawning. [ 822.897267] env[61907]: INFO nova.compute.manager [-] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Took 1.40 seconds to deallocate network for instance. [ 822.940537] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5228ddaa-0250-cae1-08e5-bcb4205a5e25, 'name': SearchDatastore_Task, 'duration_secs': 0.008238} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.942323] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13f942d9-d1c8-4f57-bf68-f5f21266d5a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.948632] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 822.948632] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c284af-924a-d2bb-90e3-1785c5c446ac" [ 822.948632] env[61907]: _type = "Task" [ 822.948632] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.959799] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c284af-924a-d2bb-90e3-1785c5c446ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.993745] env[61907]: DEBUG nova.network.neutron [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Updated VIF entry in instance network info cache for port 0f0768ac-bf89-4dc8-ba37-3e163bd551b1. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 822.994319] env[61907]: DEBUG nova.network.neutron [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Updating instance_info_cache with network_info: [{"id": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "address": "fa:16:3e:15:81:45", "network": {"id": "72104f6a-7f92-428f-9f2f-ed6a9a42fbd9", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-14201096-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c53d877a81a7422db4fdfbbc6ed37444", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d47d5e1d-e66d-4f2c-83e6-d5e78c2b767d", "external-id": "nsx-vlan-transportzone-109", "segmentation_id": 109, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0f0768ac-bf", "ovs_interfaceid": "0f0768ac-bf89-4dc8-ba37-3e163bd551b1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.060556] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.955s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.062143] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.062376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.063103] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.064492] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.577s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.066037] env[61907]: INFO nova.compute.claims [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.085598] env[61907]: INFO nova.scheduler.client.report [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocations for instance 1ef8594c-5d9a-491f-89f6-e5d7f35ce751 [ 823.127991] env[61907]: DEBUG oslo_vmware.api [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243823, 'name': PowerOnVM_Task, 'duration_secs': 1.095734} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.128151] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 823.128510] env[61907]: INFO nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Took 8.25 seconds to spawn the instance on the hypervisor. [ 823.128510] env[61907]: DEBUG nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 823.129917] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ef1447-3baa-44b3-9dfb-e603ab58a94d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.404784] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.459814] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c284af-924a-d2bb-90e3-1785c5c446ac, 'name': SearchDatastore_Task, 'duration_secs': 0.016024} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.460127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.460429] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e1f8f88b-7179-4e92-83cd-a1027ff9b0a2/e1f8f88b-7179-4e92-83cd-a1027ff9b0a2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 823.460839] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49bd70de-8465-409a-a797-a9bedf87e5ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.468169] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 823.468169] env[61907]: value = "task-1243824" [ 823.468169] env[61907]: _type = "Task" [ 823.468169] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.476248] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243824, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.497482] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] Releasing lock "refresh_cache-e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.497482] env[61907]: DEBUG nova.compute.manager [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Received event network-vif-deleted-30bd832d-f042-4736-8dea-152ff21d9739 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 823.497617] env[61907]: INFO nova.compute.manager [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Neutron deleted interface 30bd832d-f042-4736-8dea-152ff21d9739; detaching it from the instance and deleting it from the info cache [ 823.497801] env[61907]: DEBUG nova.network.neutron [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.596987] env[61907]: DEBUG oslo_concurrency.lockutils [None req-27719e45-baec-40f3-91e2-3aa3c4f4df1a tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "1ef8594c-5d9a-491f-89f6-e5d7f35ce751" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.277s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.620363] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.649911] env[61907]: INFO nova.compute.manager [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Took 34.15 seconds to build instance. [ 823.790597] env[61907]: DEBUG nova.network.neutron [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.980978] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243824, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.001077] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-263d0060-f20a-49f8-9f72-2cce0bc355ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.013395] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8365b86e-a0dc-4423-8e68-8f815406083a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.041826] env[61907]: DEBUG nova.compute.manager [req-f1a8ac86-49e3-4086-b719-0feb8751e21d req-3f2d24bb-0f22-4407-b47a-2faeb4c64af0 service nova] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Detach interface failed, port_id=30bd832d-f042-4736-8dea-152ff21d9739, reason: Instance 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 824.151631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-17d001df-df62-4917-a397-981d118e4661 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.249s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.294613] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.295010] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Instance network_info: |[{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 824.388609] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e50a21-be78-4371-8a0d-b1227edaf51c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.398997] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3f6b4b-a454-48ff-b85e-bc9ef9c88135 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.430175] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-decbeb4c-3c15-4d78-9ee8-a2c2da7a69cb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.438788] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d860a20c-62be-4e04-b6b6-3650d2b00641 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.455276] env[61907]: DEBUG nova.compute.provider_tree [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.479882] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243824, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592839} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.480184] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e1f8f88b-7179-4e92-83cd-a1027ff9b0a2/e1f8f88b-7179-4e92-83cd-a1027ff9b0a2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 824.480413] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 824.480678] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-01d422a5-c82f-46ac-9df8-8ebe0b71e1c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.488094] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 824.488094] env[61907]: value = "task-1243825" [ 824.488094] env[61907]: _type = "Task" [ 824.488094] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.498882] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243825, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.654843] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 824.876096] env[61907]: DEBUG nova.compute.manager [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received event network-changed-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 824.876312] env[61907]: DEBUG nova.compute.manager [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing instance network info cache due to event network-changed-718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 824.876545] env[61907]: DEBUG oslo_concurrency.lockutils [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] Acquiring lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.876686] env[61907]: DEBUG oslo_concurrency.lockutils [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] Acquired lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.876856] env[61907]: DEBUG nova.network.neutron [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing network info cache for port 718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 824.957941] env[61907]: DEBUG nova.scheduler.client.report [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 824.999222] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243825, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100701} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.999503] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 825.000353] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520cdf80-19b4-4551-b8e6-129e43bc9702 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.023343] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] e1f8f88b-7179-4e92-83cd-a1027ff9b0a2/e1f8f88b-7179-4e92-83cd-a1027ff9b0a2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 825.024136] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed05b5da-5714-4e8e-b14b-9688d234b8bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.044328] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 825.044328] env[61907]: value = "task-1243826" [ 825.044328] env[61907]: _type = "Task" [ 825.044328] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.053618] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243826, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.124713] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.125033] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.125247] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.125533] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.125704] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.125864] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.126111] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.126284] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.126459] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.126628] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.126808] env[61907]: DEBUG nova.virt.hardware [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.129035] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105e8624-11eb-4478-b40d-138e5ed05285 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.139638] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.139922] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.140105] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.140301] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.140453] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.140605] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.140818] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.140982] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.141170] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.141339] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.141516] env[61907]: DEBUG nova.virt.hardware [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.142923] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c00845-b909-48e4-973a-76bc7b41c47e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.149311] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-307d41f9-e2e4-40e3-8af8-6fd3d2dc764a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.166924] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 825.168104] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dce11c0-f101-4bd1-b995-b88231df68f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.175028] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:5f:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.181933] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating folder: Project (7652e98cde994af28b7bac0b81547474). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.183045] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4e8067-8368-47ec-833c-e112211d4a55 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.187386] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a13c7793-8e48-4384-b1a6-e30a25b43923 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.194221] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 825.194395] env[61907]: ERROR oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk due to incomplete transfer. [ 825.205553] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-64bdecc1-87ad-438a-91cb-9b307e1da51a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.207574] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 825.213338] env[61907]: DEBUG oslo.service.loopingcall [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.214604] env[61907]: DEBUG nova.compute.manager [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Received event network-changed-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 825.214817] env[61907]: DEBUG nova.compute.manager [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Refreshing instance network info cache due to event network-changed-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 825.215075] env[61907]: DEBUG oslo_concurrency.lockutils [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.215235] env[61907]: DEBUG oslo_concurrency.lockutils [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.215402] env[61907]: DEBUG nova.network.neutron [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Refreshing network info cache for port f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.217345] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.219073] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.219358] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created folder: Project (7652e98cde994af28b7bac0b81547474) in parent group-v268168. [ 825.219534] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating folder: Instances. Parent ref: group-v268240. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 825.220360] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-343469ee-926a-465e-bae9-35eef1ff1f4f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.232673] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-61b3b124-cc83-40df-a8cc-bf901fae6853 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.240033] env[61907]: DEBUG oslo_vmware.rw_handles [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b78575-30d8-ee99-5f57-77b3c4adf7ff/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 825.240033] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Uploaded image 39c1a265-6dd8-4c56-ac69-1ca1342620d3 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 825.240033] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 825.240313] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d60a5da9-a3f7-4259-bde1-4e859d67a925 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.245591] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.245591] env[61907]: value = "task-1243828" [ 825.245591] env[61907]: _type = "Task" [ 825.245591] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.251238] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 825.251238] env[61907]: value = "task-1243830" [ 825.251238] env[61907]: _type = "Task" [ 825.251238] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.251803] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created folder: Instances in parent group-v268240. [ 825.252050] env[61907]: DEBUG oslo.service.loopingcall [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.253915] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 825.255713] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-85a9ae10-d381-4f27-a727-306c54e6ca5f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.275587] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243828, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.282244] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243830, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.284602] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.284602] env[61907]: value = "task-1243831" [ 825.284602] env[61907]: _type = "Task" [ 825.284602] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.295281] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243831, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.463295] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.399s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.463904] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 825.466834] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.586s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.467125] env[61907]: DEBUG nova.objects.instance [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'resources' on Instance uuid 3c9ee76a-464d-437e-9b73-862f1da24f7e {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.557223] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243826, 'name': ReconfigVM_Task, 'duration_secs': 0.290726} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.557573] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Reconfigured VM instance instance-0000003e to attach disk [datastore2] e1f8f88b-7179-4e92-83cd-a1027ff9b0a2/e1f8f88b-7179-4e92-83cd-a1027ff9b0a2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 825.558318] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3d998f71-507c-481f-8f4e-249ed79b4e67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.569795] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 825.569795] env[61907]: value = "task-1243832" [ 825.569795] env[61907]: _type = "Task" [ 825.569795] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.578617] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243832, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.756785] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243828, 'name': CreateVM_Task, 'duration_secs': 0.414837} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.757770] env[61907]: DEBUG nova.network.neutron [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updated VIF entry in instance network info cache for port 718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 825.758158] env[61907]: DEBUG nova.network.neutron [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.762314] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.765192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.765397] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.765732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.766414] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d6a2955-bae1-4bf7-8315-b1ece45c8c13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.774266] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243830, 'name': Destroy_Task, 'duration_secs': 0.456128} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.775052] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Destroyed the VM [ 825.775365] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 825.775646] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0037dc9c-1d6e-4f89-b58d-1bbeb5fe2df8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.779520] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 825.779520] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b2962f-0e68-f6d4-87f5-27451909085c" [ 825.779520] env[61907]: _type = "Task" [ 825.779520] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.785575] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 825.785575] env[61907]: value = "task-1243833" [ 825.785575] env[61907]: _type = "Task" [ 825.785575] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.798158] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b2962f-0e68-f6d4-87f5-27451909085c, 'name': SearchDatastore_Task, 'duration_secs': 0.010815} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.799356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.799624] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 825.799853] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.800013] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.800305] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 825.804326] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2e2499b-4d53-4aaa-a7f4-09b117247cae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.806294] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243831, 'name': CreateVM_Task, 'duration_secs': 0.421758} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.809269] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 825.809897] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243833, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.810959] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.811230] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.811558] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 825.811874] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4551a075-c99b-4103-a2c0-743657739d56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.816793] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 825.816793] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5230cfbd-3660-f7b5-ff1a-5573f7d31bc4" [ 825.816793] env[61907]: _type = "Task" [ 825.816793] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.817926] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 825.818127] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 825.822050] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11f1468f-107f-4941-90f8-309ac5c6dae0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.829651] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "5c908f9c-7efb-4588-b6bf-70babb173767" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.829948] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.830196] env[61907]: INFO nova.compute.manager [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Rebooting instance [ 825.831496] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5230cfbd-3660-f7b5-ff1a-5573f7d31bc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.839164] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 825.839164] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523b82f9-fd03-5d02-6cb0-71e4246c6df5" [ 825.839164] env[61907]: _type = "Task" [ 825.839164] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.849281] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523b82f9-fd03-5d02-6cb0-71e4246c6df5, 'name': SearchDatastore_Task, 'duration_secs': 0.00953} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.850115] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56324df8-15b1-47b7-848d-e0300203d3f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.855997] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 825.855997] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521b5813-ecae-6482-8a91-b19c4ffec39b" [ 825.855997] env[61907]: _type = "Task" [ 825.855997] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.871854] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521b5813-ecae-6482-8a91-b19c4ffec39b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.970461] env[61907]: DEBUG nova.objects.instance [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'numa_topology' on Instance uuid 3c9ee76a-464d-437e-9b73-862f1da24f7e {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.972467] env[61907]: DEBUG nova.compute.utils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.973906] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 825.974099] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.037668] env[61907]: DEBUG nova.policy [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e8057170d8a41d2a49bf4485c352e3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89e4f90a5fe44853a926ceba2f5150dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 826.081668] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243832, 'name': Rename_Task, 'duration_secs': 0.203661} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.082215] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 826.082552] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7186207e-26bb-4227-af2e-3e0ab22fdf16 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.090136] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 826.090136] env[61907]: value = "task-1243834" [ 826.090136] env[61907]: _type = "Task" [ 826.090136] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.099499] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243834, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.106425] env[61907]: DEBUG nova.network.neutron [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updated VIF entry in instance network info cache for port f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 826.106533] env[61907]: DEBUG nova.network.neutron [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.265925] env[61907]: DEBUG oslo_concurrency.lockutils [req-31ce9f6b-8188-4303-a3c8-2a018ae9d9d1 req-491da9cd-b414-4d45-a97b-bf0205bb549c service nova] Releasing lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.301429] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243833, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.332110] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5230cfbd-3660-f7b5-ff1a-5573f7d31bc4, 'name': SearchDatastore_Task, 'duration_secs': 0.019447} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.333216] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.333216] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.333216] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.364762] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.364951] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquired lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.365813] env[61907]: DEBUG nova.network.neutron [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.373013] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521b5813-ecae-6482-8a91-b19c4ffec39b, 'name': SearchDatastore_Task, 'duration_secs': 0.027565} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.373410] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.373837] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 826.374765] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.375151] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.375394] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-19f120bc-821e-41ea-bf7c-b2be6d110869 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.377669] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-628774bc-1bb7-4977-8144-4102e3bf9ab6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.385784] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 826.385784] env[61907]: value = "task-1243835" [ 826.385784] env[61907]: _type = "Task" [ 826.385784] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.391171] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.391522] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 826.392905] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5a33126-67dd-4d5f-88b7-916f2ef8b91d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.401238] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243835, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.405401] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 826.405401] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522385a7-b590-a8bd-a3e4-101247541c97" [ 826.405401] env[61907]: _type = "Task" [ 826.405401] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.414510] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522385a7-b590-a8bd-a3e4-101247541c97, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.451529] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "ce864af6-aef6-4044-be64-8440ba175438" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.451921] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.452176] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "ce864af6-aef6-4044-be64-8440ba175438-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.452388] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.452589] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.455389] env[61907]: INFO nova.compute.manager [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Terminating instance [ 826.460265] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Successfully created port: 9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 826.474858] env[61907]: DEBUG nova.objects.base [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Object Instance<3c9ee76a-464d-437e-9b73-862f1da24f7e> lazy-loaded attributes: resources,numa_topology {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 826.478568] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 826.605642] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243834, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.612638] env[61907]: DEBUG oslo_concurrency.lockutils [req-b899057c-0fd8-4771-b8f8-a16c32f5d6c5 req-9a09b3c9-d40a-4577-b425-b17acb0433df service nova] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.803857] env[61907]: DEBUG oslo_vmware.api [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243833, 'name': RemoveSnapshot_Task, 'duration_secs': 0.653787} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.804409] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 826.804726] env[61907]: INFO nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Took 13.59 seconds to snapshot the instance on the hypervisor. [ 826.900554] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243835, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.919113] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522385a7-b590-a8bd-a3e4-101247541c97, 'name': SearchDatastore_Task, 'duration_secs': 0.014935} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.927100] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42109a60-90bb-479b-ba07-035e610b0ad6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.935288] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 826.935288] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526f3c71-b401-c059-188d-7431b4077341" [ 826.935288] env[61907]: _type = "Task" [ 826.935288] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.945933] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526f3c71-b401-c059-188d-7431b4077341, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.960239] env[61907]: DEBUG nova.compute.manager [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 826.960443] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 826.961382] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1526d0f1-28d7-4c1e-9f14-db0f21aa8aaf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.973144] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 826.973144] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b01f9698-effe-4b59-b1d2-807462eec188 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.026612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ff533f2-8ece-4794-a0e1-eea7e2cde474 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.035724] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6a28dc-5cb6-4db1-a575-581b30ed2c70 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.075403] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a07a3bb-b06b-43c4-9c8b-34bf98e3b029 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.085028] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bedb88b3-3adf-4eba-900b-edaf42d292b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.093175] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 827.093175] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 827.093175] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleting the datastore file [datastore1] ce864af6-aef6-4044-be64-8440ba175438 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.093175] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f6def2a-17af-4cb2-a3ce-d8d0a994c31e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.104844] env[61907]: DEBUG nova.compute.provider_tree [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.115248] env[61907]: DEBUG oslo_vmware.api [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 827.115248] env[61907]: value = "task-1243837" [ 827.115248] env[61907]: _type = "Task" [ 827.115248] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.118295] env[61907]: DEBUG oslo_vmware.api [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243834, 'name': PowerOnVM_Task, 'duration_secs': 0.533086} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.120777] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 827.124643] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 827.124861] env[61907]: INFO nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Took 7.71 seconds to spawn the instance on the hypervisor. [ 827.125064] env[61907]: DEBUG nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 827.126241] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24bd2057-ae04-4a82-a249-d8a0caea4355 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.130842] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edaf5101-3375-4642-9f4a-0050565e6e77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.136837] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "e52648ac-e068-4a63-baa0-a1c34df52197" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.137118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.141174] env[61907]: DEBUG oslo_vmware.api [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.143777] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 827.143940] env[61907]: ERROR oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk due to incomplete transfer. [ 827.146595] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0c6d1e46-e90c-4fcb-8d5f-4bc2e2835502 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.155552] env[61907]: DEBUG oslo_vmware.rw_handles [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5207edf9-0669-a79b-d799-cf7e16e8437b/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 827.155552] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Uploaded image c525c73f-cabe-4077-be2c-3ee109d7c7c8 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 827.159019] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 827.159019] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dd0c8edc-0ce0-4eb9-ac09-505602d333ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.164650] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 827.164650] env[61907]: value = "task-1243838" [ 827.164650] env[61907]: _type = "Task" [ 827.164650] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.177156] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243838, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.226306] env[61907]: DEBUG nova.network.neutron [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.308986] env[61907]: DEBUG nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance disappeared during snapshot {{(pid=61907) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4594}} [ 827.323394] env[61907]: DEBUG nova.compute.manager [None req-b4d0f52b-fc7d-41d0-bf0d-4b222858a249 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image not found during clean up 39c1a265-6dd8-4c56-ac69-1ca1342620d3 {{(pid=61907) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4600}} [ 827.399724] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243835, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.576988} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.400391] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 827.400620] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 827.400887] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83dc2e74-12d0-478f-916b-20f9cef99068 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.406956] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 827.406956] env[61907]: value = "task-1243839" [ 827.406956] env[61907]: _type = "Task" [ 827.406956] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.414921] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243839, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.445772] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526f3c71-b401-c059-188d-7431b4077341, 'name': SearchDatastore_Task, 'duration_secs': 0.015167} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.446110] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.446392] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 827.446684] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4c06c403-e0ef-4be3-81be-05576ba922a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.452983] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 827.452983] env[61907]: value = "task-1243840" [ 827.452983] env[61907]: _type = "Task" [ 827.452983] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.460919] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.490934] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 827.513714] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 827.514053] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 827.514237] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 827.514430] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 827.514583] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 827.514740] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 827.514959] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 827.515141] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 827.515316] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 827.515483] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 827.515664] env[61907]: DEBUG nova.virt.hardware [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 827.516548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf94ff6-2663-44ec-9ce6-efe296421076 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.525035] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7305e03-c69d-4b53-a0fe-380f8d271656 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.614471] env[61907]: DEBUG nova.scheduler.client.report [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 827.627456] env[61907]: DEBUG oslo_vmware.api [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243837, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.656059] env[61907]: INFO nova.compute.manager [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Took 35.49 seconds to build instance. [ 827.677407] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243838, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.729168] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Releasing lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.916475] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243839, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.274836} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.916778] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 827.917566] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c044f17-0518-4538-89cc-4b026af99bb5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.937242] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 827.937578] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cdb1afe8-731c-459d-a98f-426d8ecaac8d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.959954] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 827.959954] env[61907]: value = "task-1243841" [ 827.959954] env[61907]: _type = "Task" [ 827.959954] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.966313] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243840, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.970645] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243841, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.120062] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.653s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.123833] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.185s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.125451] env[61907]: INFO nova.compute.claims [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.137167] env[61907]: DEBUG oslo_vmware.api [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243837, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.82133} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.137525] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 828.137768] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 828.137976] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 828.138175] env[61907]: INFO nova.compute.manager [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: ce864af6-aef6-4044-be64-8440ba175438] Took 1.18 seconds to destroy the instance on the hypervisor. [ 828.138425] env[61907]: DEBUG oslo.service.loopingcall [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.138660] env[61907]: DEBUG nova.compute.manager [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 828.138752] env[61907]: DEBUG nova.network.neutron [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 828.158304] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fe204337-7ce1-4673-a255-e2e67f761f82 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.038s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.177341] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243838, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.183204] env[61907]: DEBUG nova.compute.manager [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Received event network-vif-plugged-9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 828.184337] env[61907]: DEBUG oslo_concurrency.lockutils [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] Acquiring lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.184337] env[61907]: DEBUG oslo_concurrency.lockutils [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.184337] env[61907]: DEBUG oslo_concurrency.lockutils [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.185069] env[61907]: DEBUG nova.compute.manager [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] No waiting events found dispatching network-vif-plugged-9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 828.185069] env[61907]: WARNING nova.compute.manager [req-7abe857d-c162-4ed1-92e6-fd306713b473 req-538cb99f-f745-443b-a286-c063baa4706c service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Received unexpected event network-vif-plugged-9a182520-96f9-434d-844a-fe3e6cc88fbc for instance with vm_state building and task_state spawning. [ 828.235222] env[61907]: DEBUG nova.compute.manager [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 828.235984] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae645e89-31e8-4c6f-920e-ff46b9a771e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.265143] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Successfully updated port: 9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.465330] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243840, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.481905] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243841, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.574692] env[61907]: DEBUG nova.compute.manager [req-51fe4610-bf1f-4973-b4d5-01c0c9b9541d req-4c27a1f6-0565-4919-ac40-4ab4908414ee service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Received event network-vif-deleted-da5f6c5a-0d2b-4fc6-8c90-1bfc00457236 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 828.574925] env[61907]: INFO nova.compute.manager [req-51fe4610-bf1f-4973-b4d5-01c0c9b9541d req-4c27a1f6-0565-4919-ac40-4ab4908414ee service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Neutron deleted interface da5f6c5a-0d2b-4fc6-8c90-1bfc00457236; detaching it from the instance and deleting it from the info cache [ 828.575137] env[61907]: DEBUG nova.network.neutron [req-51fe4610-bf1f-4973-b4d5-01c0c9b9541d req-4c27a1f6-0565-4919-ac40-4ab4908414ee service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.628417] env[61907]: DEBUG oslo_concurrency.lockutils [None req-98a35682-e116-46d4-9fd2-316d88e65317 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 44.103s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.629463] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 23.024s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.629793] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.630054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.630244] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.633689] env[61907]: INFO nova.compute.manager [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Terminating instance [ 828.660855] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 828.675481] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243838, 'name': Destroy_Task, 'duration_secs': 1.364515} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.675768] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Destroyed the VM [ 828.676063] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 828.676342] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-19d37b3d-a7fd-44b6-a408-3c0dd85d8ee1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.682423] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 828.682423] env[61907]: value = "task-1243842" [ 828.682423] env[61907]: _type = "Task" [ 828.682423] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.690368] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243842, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.767392] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.767764] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.767764] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.964704] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243840, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.186612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.965653] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 828.965653] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.968721] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-956284e2-1e28-4821-9e5b-3764f5081641 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.975448] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243841, 'name': ReconfigVM_Task, 'duration_secs': 0.950298} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.976628] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Reconfigured VM instance instance-0000003c to attach disk [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182/e5d4890e-be1b-4b28-9721-89ef678c1182.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 828.977533] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 828.977533] env[61907]: value = "task-1243843" [ 828.977533] env[61907]: _type = "Task" [ 828.977533] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.977533] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-966f635f-c029-44b9-be24-22b632ce6c2c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.989679] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.991101] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 828.991101] env[61907]: value = "task-1243844" [ 828.991101] env[61907]: _type = "Task" [ 828.991101] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.998900] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243844, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.014486] env[61907]: DEBUG nova.network.neutron [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.077661] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-84a759b9-dec8-48a2-b643-7ea8b07f1638 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.087142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c717ff1f-4026-4b58-8f6d-258b21904c15 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.114369] env[61907]: DEBUG nova.compute.manager [req-51fe4610-bf1f-4973-b4d5-01c0c9b9541d req-4c27a1f6-0565-4919-ac40-4ab4908414ee service nova] [instance: ce864af6-aef6-4044-be64-8440ba175438] Detach interface failed, port_id=da5f6c5a-0d2b-4fc6-8c90-1bfc00457236, reason: Instance ce864af6-aef6-4044-be64-8440ba175438 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 829.140062] env[61907]: DEBUG nova.compute.manager [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 829.140347] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.140652] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e36c80bd-56bf-49c4-898f-2d49db5a90df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.149918] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8156cec-26b8-4821-802f-733fee097b21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.179301] env[61907]: WARNING nova.virt.vmwareapi.vmops [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c9ee76a-464d-437e-9b73-862f1da24f7e could not be found. [ 829.179530] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.179746] env[61907]: INFO nova.compute.manager [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 829.180048] env[61907]: DEBUG oslo.service.loopingcall [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.180969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.183358] env[61907]: DEBUG nova.compute.manager [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 829.183460] env[61907]: DEBUG nova.network.neutron [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.193592] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243842, 'name': RemoveSnapshot_Task, 'duration_secs': 0.338176} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.193873] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 829.194183] env[61907]: DEBUG nova.compute.manager [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 829.194986] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33658c81-8315-47a4-80f6-d0c2608bb301 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.253926] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6872253c-b87d-4204-a6ec-5321f98de8fe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.265457] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Doing hard reboot of VM {{(pid=61907) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 829.265779] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-0f78366e-abe8-479b-9a60-fe7a3b3dc351 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.274488] env[61907]: DEBUG oslo_vmware.api [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 829.274488] env[61907]: value = "task-1243845" [ 829.274488] env[61907]: _type = "Task" [ 829.274488] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.285049] env[61907]: DEBUG oslo_vmware.api [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243845, 'name': ResetVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.312583] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.475442] env[61907]: DEBUG nova.network.neutron [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.487786] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130179} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.489046] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.489671] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd57b960-b191-460e-8205-f1a1b1b60289 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.493648] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1476fbf2-2333-4870-adf7-27e131ea49be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.514879] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d800a876-ebc4-43de-9f1d-7938b7fea1ca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.526721] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.527068] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243844, 'name': Rename_Task, 'duration_secs': 0.242133} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.527467] env[61907]: INFO nova.compute.manager [-] [instance: ce864af6-aef6-4044-be64-8440ba175438] Took 1.39 seconds to deallocate network for instance. [ 829.528059] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3557530-2b54-4342-bd22-3d2b815f10a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.541824] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 829.543877] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef110978-666c-41d8-b284-afa107d37adf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.577080] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57b27229-cc33-4d71-b671-b8085e9b416c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.580008] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 829.580008] env[61907]: value = "task-1243846" [ 829.580008] env[61907]: _type = "Task" [ 829.580008] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.581707] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 829.581707] env[61907]: value = "task-1243847" [ 829.581707] env[61907]: _type = "Task" [ 829.581707] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.593732] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ed3ba7-46c8-4f2c-874e-7c0acc2c2b27 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.604257] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.604557] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243847, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.615703] env[61907]: DEBUG nova.compute.provider_tree [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.710815] env[61907]: INFO nova.compute.manager [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Shelve offloading [ 829.785866] env[61907]: DEBUG oslo_vmware.api [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243845, 'name': ResetVM_Task, 'duration_secs': 0.095448} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.786223] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Did hard reboot of VM {{(pid=61907) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 829.786427] env[61907]: DEBUG nova.compute.manager [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 829.787538] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74241a71-f25c-4d49-8bae-0043481ccd1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.909966] env[61907]: DEBUG nova.network.neutron [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.977624] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.978077] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Instance network_info: |[{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 829.978746] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:80:83:7b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9a182520-96f9-434d-844a-fe3e6cc88fbc', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.986915] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating folder: Project (89e4f90a5fe44853a926ceba2f5150dd). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 829.987502] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2c55d91-d288-4957-91b7-2fdc93a971cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.999525] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created folder: Project (89e4f90a5fe44853a926ceba2f5150dd) in parent group-v268168. [ 829.999795] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating folder: Instances. Parent ref: group-v268244. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.000110] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9cafd125-e4e4-4bee-a569-ec821e04a159 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.009504] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created folder: Instances in parent group-v268244. [ 830.010909] env[61907]: DEBUG oslo.service.loopingcall [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.010909] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.010909] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-192f2f35-080f-4d77-89c2-507c2c8c2af3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.029195] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.029195] env[61907]: value = "task-1243850" [ 830.029195] env[61907]: _type = "Task" [ 830.029195] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.036296] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243850, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.076147] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.093598] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243846, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.096948] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243847, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.119251] env[61907]: DEBUG nova.scheduler.client.report [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 830.215773] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.216167] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7611f4cb-f4d6-49dd-922f-243af0b7ca3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.223886] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 830.223886] env[61907]: value = "task-1243851" [ 830.223886] env[61907]: _type = "Task" [ 830.223886] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.232365] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.291646] env[61907]: DEBUG nova.compute.manager [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Received event network-changed-9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 830.292111] env[61907]: DEBUG nova.compute.manager [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Refreshing instance network info cache due to event network-changed-9a182520-96f9-434d-844a-fe3e6cc88fbc. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 830.292622] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] Acquiring lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.294284] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] Acquired lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.294284] env[61907]: DEBUG nova.network.neutron [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Refreshing network info cache for port 9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 830.301520] env[61907]: DEBUG oslo_concurrency.lockutils [None req-358c1f7c-01a5-4646-8c87-070fb7946da5 tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.471s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.411937] env[61907]: INFO nova.compute.manager [-] [instance: 3c9ee76a-464d-437e-9b73-862f1da24f7e] Took 1.23 seconds to deallocate network for instance. [ 830.539521] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243850, 'name': CreateVM_Task, 'duration_secs': 0.366994} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.539689] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 830.540417] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.540588] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.541061] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.541445] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a23b957-bb04-41dd-a4d0-ab90df480e9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.545934] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 830.545934] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d5549-3b64-9634-2a79-185c587d1db0" [ 830.545934] env[61907]: _type = "Task" [ 830.545934] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.553311] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d5549-3b64-9634-2a79-185c587d1db0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.593745] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243846, 'name': ReconfigVM_Task, 'duration_secs': 0.923577} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.596837] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 830.597523] env[61907]: DEBUG oslo_vmware.api [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243847, 'name': PowerOnVM_Task, 'duration_secs': 0.787231} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.597737] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72631835-e5dc-41b8-b1a1-fac047e46c78 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.599250] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 830.599468] env[61907]: DEBUG nova.compute.manager [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.600397] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f99a8c-3ce5-4473-b5c1-517012f67310 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.608898] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 830.608898] env[61907]: value = "task-1243852" [ 830.608898] env[61907]: _type = "Task" [ 830.608898] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.617105] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243852, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.624516] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.501s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.625110] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 830.627366] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.475s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.628855] env[61907]: INFO nova.compute.claims [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.733973] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 830.734356] env[61907]: DEBUG nova.compute.manager [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 830.735244] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8b193e-6e5a-4134-928c-3b51d3f5ae35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.740708] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.740894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.741072] env[61907]: DEBUG nova.network.neutron [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 831.058281] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d5549-3b64-9634-2a79-185c587d1db0, 'name': SearchDatastore_Task, 'duration_secs': 0.043858} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.058617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.058859] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 831.059112] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 831.059254] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.059444] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 831.059766] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea5e0396-2abc-4b12-82e6-d7e8de73e566 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.067954] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 831.068246] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 831.068880] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d2fb343-a27e-4848-b351-a0609e17fbf3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.074362] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 831.074362] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52461943-7871-21d9-ba73-8039747bc592" [ 831.074362] env[61907]: _type = "Task" [ 831.074362] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.081708] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52461943-7871-21d9-ba73-8039747bc592, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.092278] env[61907]: DEBUG nova.network.neutron [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updated VIF entry in instance network info cache for port 9a182520-96f9-434d-844a-fe3e6cc88fbc. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 831.092622] env[61907]: DEBUG nova.network.neutron [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.116707] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.124098] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "5c908f9c-7efb-4588-b6bf-70babb173767" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.124351] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.124577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.124804] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.124988] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.127397] env[61907]: INFO nova.compute.manager [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Terminating instance [ 831.134111] env[61907]: DEBUG nova.compute.utils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.139042] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243852, 'name': Rename_Task, 'duration_secs': 0.313358} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.139042] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 831.139042] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.142167] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.142882] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7dd06e4b-a326-43e5-bb31-0ff7d7169399 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.150814] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 831.150814] env[61907]: value = "task-1243853" [ 831.150814] env[61907]: _type = "Task" [ 831.150814] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.160833] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243853, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.182215] env[61907]: DEBUG nova.policy [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e8057170d8a41d2a49bf4485c352e3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89e4f90a5fe44853a926ceba2f5150dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.345348] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.345691] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.345916] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.346267] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.346511] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.348676] env[61907]: INFO nova.compute.manager [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Terminating instance [ 831.437602] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5cb00621-a7b3-450a-ba05-dd4be58367cd tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "3c9ee76a-464d-437e-9b73-862f1da24f7e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.808s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.499385] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Successfully created port: 1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.551669] env[61907]: DEBUG nova.network.neutron [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.584366] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52461943-7871-21d9-ba73-8039747bc592, 'name': SearchDatastore_Task, 'duration_secs': 0.010277} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.585545] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d3f271d-8d8a-4f14-b89e-ffd2cc2c71f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.590708] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 831.590708] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb0f60-4af3-0c6f-6f21-f91ceaa69604" [ 831.590708] env[61907]: _type = "Task" [ 831.590708] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.595510] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ef345b4-f7d4-487d-9020-517c270faec2 req-dbd63d64-c0fb-4c3a-86ac-ee5d8b463ecc service nova] Releasing lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.600922] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb0f60-4af3-0c6f-6f21-f91ceaa69604, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.607427] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "e5d4890e-be1b-4b28-9721-89ef678c1182" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.607678] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.607892] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "e5d4890e-be1b-4b28-9721-89ef678c1182-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.608102] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.608290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.610422] env[61907]: INFO nova.compute.manager [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Terminating instance [ 831.636509] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 831.644018] env[61907]: DEBUG nova.compute.manager [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 831.644269] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.648445] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4771ad-035a-47e6-8af3-0a340b12ae85 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.661027] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243853, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.663194] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.663451] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6514eed-cd56-4ad0-b80d-cf04ebf02b08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.669440] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 831.669440] env[61907]: value = "task-1243854" [ 831.669440] env[61907]: _type = "Task" [ 831.669440] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.681476] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243854, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.853225] env[61907]: DEBUG nova.compute.manager [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 831.853525] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 831.854970] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6064f348-bcb9-4f7a-ae44-4a5ff4892b4a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.862767] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 831.863093] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10917063-0236-44fd-85d9-2d8c8fbfbd01 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.871908] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 831.871908] env[61907]: value = "task-1243855" [ 831.871908] env[61907]: _type = "Task" [ 831.871908] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.882082] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243855, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.988275] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7dd339-809a-4631-bc43-2516fd9f43c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.996225] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-097aa813-068b-4fe1-a1f5-672f55f3c15a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.029750] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a68e421-887a-4a7a-ba8c-8b07cf53fd74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.037594] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89003cee-41b5-49e4-a62c-922baa0214c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.051773] env[61907]: DEBUG nova.compute.provider_tree [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.053901] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.105052] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb0f60-4af3-0c6f-6f21-f91ceaa69604, 'name': SearchDatastore_Task, 'duration_secs': 0.011318} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.105375] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.105643] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/2ab93ab9-bf69-4525-8df6-eef83dd24bc1.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 832.106314] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3207e254-4d80-4ded-a9ac-6e53f311a55d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.117191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "refresh_cache-e5d4890e-be1b-4b28-9721-89ef678c1182" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.117191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquired lock "refresh_cache-e5d4890e-be1b-4b28-9721-89ef678c1182" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.117191] env[61907]: DEBUG nova.network.neutron [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.117191] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 832.117191] env[61907]: value = "task-1243856" [ 832.117191] env[61907]: _type = "Task" [ 832.117191] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.125495] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243856, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.169128] env[61907]: DEBUG oslo_vmware.api [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243853, 'name': PowerOnVM_Task, 'duration_secs': 0.820562} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.169183] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.169398] env[61907]: INFO nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Took 10.06 seconds to spawn the instance on the hypervisor. [ 832.169578] env[61907]: DEBUG nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 832.170436] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f604b96-514e-488e-8748-c8dc9b1a8c2c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.182189] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243854, 'name': PowerOffVM_Task, 'duration_secs': 0.242618} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.183928] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.184413] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.187150] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eedf7398-3544-4b57-b751-42571df601cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.276551] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.276551] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.276551] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Deleting the datastore file [datastore2] 5c908f9c-7efb-4588-b6bf-70babb173767 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.276551] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-146daef9-0ea0-46b1-ac78-d6a2f13eab9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.283022] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for the task: (returnval){ [ 832.283022] env[61907]: value = "task-1243858" [ 832.283022] env[61907]: _type = "Task" [ 832.283022] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.292138] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243858, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.383600] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243855, 'name': PowerOffVM_Task, 'duration_secs': 0.286774} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.383600] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 832.383600] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.383600] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f41948e7-3247-4567-a679-2ed8e21d3fe8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.458282] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.458282] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.458535] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Deleting the datastore file [datastore2] e1f8f88b-7179-4e92-83cd-a1027ff9b0a2 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.458682] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-edc29f8f-fe34-410b-acb0-a16522615394 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.466551] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for the task: (returnval){ [ 832.466551] env[61907]: value = "task-1243860" [ 832.466551] env[61907]: _type = "Task" [ 832.466551] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.476228] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243860, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.514905] env[61907]: DEBUG nova.compute.manager [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received event network-changed-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 832.514905] env[61907]: DEBUG nova.compute.manager [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing instance network info cache due to event network-changed-718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 832.514905] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] Acquiring lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.514905] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] Acquired lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.515314] env[61907]: DEBUG nova.network.neutron [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Refreshing network info cache for port 718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.555578] env[61907]: DEBUG nova.scheduler.client.report [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 832.589227] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 832.590330] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe65bd4-6516-43cf-9ca9-dcbc7f7bdb4f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.599987] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 832.600400] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-54eb4d64-ed15-47ba-9c10-dc7efcd07a04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.631355] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243856, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.645476] env[61907]: DEBUG nova.compute.manager [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-vif-unplugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 832.647542] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.647542] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.647542] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.647542] env[61907]: DEBUG nova.compute.manager [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] No waiting events found dispatching network-vif-unplugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 832.647542] env[61907]: WARNING nova.compute.manager [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received unexpected event network-vif-unplugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 for instance with vm_state shelved and task_state shelving_offloading. [ 832.648200] env[61907]: DEBUG nova.compute.manager [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 832.648200] env[61907]: DEBUG nova.compute.manager [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing instance network info cache due to event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 832.648200] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.648298] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.649394] env[61907]: DEBUG nova.network.neutron [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.650998] env[61907]: DEBUG nova.network.neutron [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.655037] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 832.686029] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 832.686029] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 832.686029] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleting the datastore file [datastore1] 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.686292] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9fea10f-3ff1-4c66-8498-7b475a9694cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.697539] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.697872] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.698059] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.698274] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.698433] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.698592] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.698899] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.698980] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.699194] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.699374] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.699811] env[61907]: DEBUG nova.virt.hardware [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.703043] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c79bcefb-1447-4419-92b9-86a8d48d041f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.715074] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.715408] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.715741] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 832.715741] env[61907]: value = "task-1243862" [ 832.715741] env[61907]: _type = "Task" [ 832.715741] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.716266] env[61907]: INFO nova.compute.manager [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Took 37.46 seconds to build instance. [ 832.729185] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98110345-411a-4371-815c-ee46524efc8d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.739033] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.768699] env[61907]: DEBUG nova.network.neutron [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.794482] env[61907]: DEBUG oslo_vmware.api [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Task: {'id': task-1243858, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.45385} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.794831] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.795499] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.795499] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.795499] env[61907]: INFO nova.compute.manager [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Took 1.15 seconds to destroy the instance on the hypervisor. [ 832.795726] env[61907]: DEBUG oslo.service.loopingcall [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.795862] env[61907]: DEBUG nova.compute.manager [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 832.795957] env[61907]: DEBUG nova.network.neutron [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 832.977866] env[61907]: DEBUG oslo_vmware.api [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Task: {'id': task-1243860, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.394676} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.978539] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.978798] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 832.979283] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 832.979575] env[61907]: INFO nova.compute.manager [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 832.979907] env[61907]: DEBUG oslo.service.loopingcall [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 832.980139] env[61907]: DEBUG nova.compute.manager [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 832.980258] env[61907]: DEBUG nova.network.neutron [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.053737] env[61907]: DEBUG nova.compute.manager [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Received event network-vif-plugged-1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 833.053811] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] Acquiring lock "23526967-cd8f-4581-b9c5-1c270d385163-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.054297] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] Lock "23526967-cd8f-4581-b9c5-1c270d385163-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.054558] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] Lock "23526967-cd8f-4581-b9c5-1c270d385163-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.054775] env[61907]: DEBUG nova.compute.manager [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] No waiting events found dispatching network-vif-plugged-1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 833.055183] env[61907]: WARNING nova.compute.manager [req-f4bfed12-bed1-4440-99de-f720671f4956 req-c85ce60f-cf4c-4586-bcf4-2f8c3e1c87a7 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Received unexpected event network-vif-plugged-1773082b-e928-4ca3-ab1c-515606cb7e12 for instance with vm_state building and task_state spawning. [ 833.060835] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.433s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.061457] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 833.064090] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.197s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.064356] env[61907]: DEBUG nova.objects.instance [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lazy-loading 'resources' on Instance uuid 5b20f363-dfad-4c2b-b757-4da4e5bd0b99 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 833.124499] env[61907]: DEBUG nova.network.neutron [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updated VIF entry in instance network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.124499] env[61907]: DEBUG nova.network.neutron [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap40a79714-91", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.130902] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243856, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.631219} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.131198] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/2ab93ab9-bf69-4525-8df6-eef83dd24bc1.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 833.131427] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 833.131695] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-776bac2b-3380-4e0f-bd78-aac910d4ac81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.140525] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 833.140525] env[61907]: value = "task-1243863" [ 833.140525] env[61907]: _type = "Task" [ 833.140525] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.155620] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243863, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.222321] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1600d67-e93e-457f-a1d9-fe73f7937714 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.412s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.228643] env[61907]: DEBUG oslo_vmware.api [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1243862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192016} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.228920] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.229094] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 833.229287] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 833.262327] env[61907]: INFO nova.scheduler.client.report [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted allocations for instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe [ 833.273052] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Releasing lock "refresh_cache-e5d4890e-be1b-4b28-9721-89ef678c1182" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.273052] env[61907]: DEBUG nova.compute.manager [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 833.273052] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 833.274013] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83cb0170-e3cd-4008-96a4-58961df409d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.282416] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 833.282690] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3b802107-3671-413c-99d8-45b770f95fb5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.290155] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 833.290155] env[61907]: value = "task-1243864" [ 833.290155] env[61907]: _type = "Task" [ 833.290155] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.300266] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243864, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.517068] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Successfully updated port: 1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.567792] env[61907]: DEBUG nova.compute.utils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.572295] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 833.572509] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.626378] env[61907]: DEBUG nova.network.neutron [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updated VIF entry in instance network info cache for port 718419f3-2250-41d9-88ef-464d9384887a. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 833.626637] env[61907]: DEBUG nova.network.neutron [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [{"id": "718419f3-2250-41d9-88ef-464d9384887a", "address": "fa:16:3e:d8:10:52", "network": {"id": "40d280c5-52a7-460e-87a4-7eaf4d0fd635", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-779409895-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2c808b1baf1842fbb9e2d28f0031e4d7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "62d6a386-ffdb-4232-83f3-cb21c5e59e85", "external-id": "nsx-vlan-transportzone-950", "segmentation_id": 950, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap718419f3-22", "ovs_interfaceid": "718419f3-2250-41d9-88ef-464d9384887a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.628421] env[61907]: DEBUG oslo_concurrency.lockutils [req-11e578ab-8462-45b5-8bfe-9330037b55c6 req-e8ba5e06-996f-4c04-8799-84d36169baed service nova] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.655738] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243863, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072419} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.655738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.655738] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5753b0b-fb96-4bea-8b21-c6ac2059a7e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.667444] env[61907]: DEBUG nova.policy [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15bfb880d08342938b3fc3053b8090f0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e704030091a4d7985062c2a39d4f1b5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.686308] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/2ab93ab9-bf69-4525-8df6-eef83dd24bc1.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.689611] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d945712a-e0c8-407e-add4-eab191751c41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.710595] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 833.710595] env[61907]: value = "task-1243865" [ 833.710595] env[61907]: _type = "Task" [ 833.710595] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.723529] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243865, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.725041] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 833.768961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.801165] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243864, 'name': PowerOffVM_Task, 'duration_secs': 0.186969} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.806017] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 833.806017] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 833.806017] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c128a21-ab81-406d-97d2-ca93c8de3222 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.830320] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 833.830748] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 833.831037] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Deleting the datastore file [datastore2] e5d4890e-be1b-4b28-9721-89ef678c1182 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 833.831312] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-672b8365-1f9d-4ed5-85c0-a6726f2c7314 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.841492] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for the task: (returnval){ [ 833.841492] env[61907]: value = "task-1243867" [ 833.841492] env[61907]: _type = "Task" [ 833.841492] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.849182] env[61907]: DEBUG nova.network.neutron [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.854423] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.019843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.019993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.020173] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.023619] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b631feca-d545-4c78-ac11-43224f990227 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.026854] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Successfully created port: e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.033794] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3158818e-75d0-4170-8a04-6e3999c1c241 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.069072] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754d1933-8eee-4fc8-a591-30460e29e1b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.074030] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 834.078413] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c58fe2-5f0f-4a79-9446-27286de9fed8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.091998] env[61907]: DEBUG nova.compute.provider_tree [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.130168] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea408429-2c33-4e6f-9331-2b75865be5fe req-f2218f0f-5008-45f0-bfdd-51581691c076 service nova] Releasing lock "refresh_cache-5c908f9c-7efb-4588-b6bf-70babb173767" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.164238] env[61907]: DEBUG nova.network.neutron [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.221357] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243865, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.247135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.354681] env[61907]: DEBUG oslo_vmware.api [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Task: {'id': task-1243867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090146} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.354681] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.354681] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 834.354681] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.354681] env[61907]: INFO nova.compute.manager [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Took 1.08 seconds to destroy the instance on the hypervisor. [ 834.354910] env[61907]: DEBUG oslo.service.loopingcall [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.354910] env[61907]: DEBUG nova.compute.manager [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 834.354910] env[61907]: DEBUG nova.network.neutron [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.355488] env[61907]: INFO nova.compute.manager [-] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Took 1.56 seconds to deallocate network for instance. [ 834.374768] env[61907]: DEBUG nova.network.neutron [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.560846] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.594599] env[61907]: DEBUG nova.scheduler.client.report [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 834.648143] env[61907]: DEBUG nova.compute.manager [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Received event network-changed-1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 834.648390] env[61907]: DEBUG nova.compute.manager [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Refreshing instance network info cache due to event network-changed-1773082b-e928-4ca3-ab1c-515606cb7e12. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 834.648634] env[61907]: DEBUG oslo_concurrency.lockutils [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] Acquiring lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.667151] env[61907]: INFO nova.compute.manager [-] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Took 1.69 seconds to deallocate network for instance. [ 834.723637] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243865, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.724768] env[61907]: DEBUG nova.network.neutron [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Updating instance_info_cache with network_info: [{"id": "1773082b-e928-4ca3-ab1c-515606cb7e12", "address": "fa:16:3e:0b:b9:a2", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1773082b-e9", "ovs_interfaceid": "1773082b-e928-4ca3-ab1c-515606cb7e12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.863715] env[61907]: DEBUG nova.compute.manager [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Received event network-vif-deleted-718419f3-2250-41d9-88ef-464d9384887a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 834.864038] env[61907]: DEBUG nova.compute.manager [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Received event network-changed-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 834.864147] env[61907]: DEBUG nova.compute.manager [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Refreshing instance network info cache due to event network-changed-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 834.864363] env[61907]: DEBUG oslo_concurrency.lockutils [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.864544] env[61907]: DEBUG oslo_concurrency.lockutils [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.864715] env[61907]: DEBUG nova.network.neutron [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Refreshing network info cache for port f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 834.867125] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 834.877255] env[61907]: DEBUG nova.network.neutron [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.091402] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 835.099821] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.035s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.102100] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.085s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.102377] env[61907]: DEBUG nova.objects.instance [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lazy-loading 'resources' on Instance uuid 08647e48-9e24-4d7c-a9fd-4066317a7250 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.113296] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.113590] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.113763] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.113958] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.114130] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.114285] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.114496] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.114662] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.114951] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.114986] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.115278] env[61907]: DEBUG nova.virt.hardware [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.116407] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18a8fed8-3d86-46eb-9a20-4dfd3f4c6fd6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.124704] env[61907]: INFO nova.scheduler.client.report [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted allocations for instance 5b20f363-dfad-4c2b-b757-4da4e5bd0b99 [ 835.127754] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56822a7d-3f45-410e-be19-feb9b11309ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.173818] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.221923] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243865, 'name': ReconfigVM_Task, 'duration_secs': 1.281612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.222260] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/2ab93ab9-bf69-4525-8df6-eef83dd24bc1.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.222910] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-669cc75c-e8bf-4cd3-ad8f-083e48ba23ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.227638] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.227638] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Instance network_info: |[{"id": "1773082b-e928-4ca3-ab1c-515606cb7e12", "address": "fa:16:3e:0b:b9:a2", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1773082b-e9", "ovs_interfaceid": "1773082b-e928-4ca3-ab1c-515606cb7e12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 835.227801] env[61907]: DEBUG oslo_concurrency.lockutils [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] Acquired lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.227912] env[61907]: DEBUG nova.network.neutron [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Refreshing network info cache for port 1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.229577] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:b9:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1773082b-e928-4ca3-ab1c-515606cb7e12', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.236343] env[61907]: DEBUG oslo.service.loopingcall [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.240738] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.240822] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 835.240822] env[61907]: value = "task-1243868" [ 835.240822] env[61907]: _type = "Task" [ 835.240822] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.241270] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d84cd8de-96e0-45d6-962e-517762992805 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.265821] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243868, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.267161] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.267161] env[61907]: value = "task-1243869" [ 835.267161] env[61907]: _type = "Task" [ 835.267161] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.275442] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243869, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.382352] env[61907]: INFO nova.compute.manager [-] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Took 1.03 seconds to deallocate network for instance. [ 835.564900] env[61907]: DEBUG nova.compute.manager [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-vif-plugged-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 835.565619] env[61907]: DEBUG oslo_concurrency.lockutils [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] Acquiring lock "da490943-511b-4776-8f16-4f51c3b055c4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.566022] env[61907]: DEBUG oslo_concurrency.lockutils [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] Lock "da490943-511b-4776-8f16-4f51c3b055c4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.566319] env[61907]: DEBUG oslo_concurrency.lockutils [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] Lock "da490943-511b-4776-8f16-4f51c3b055c4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.566597] env[61907]: DEBUG nova.compute.manager [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] No waiting events found dispatching network-vif-plugged-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 835.566868] env[61907]: WARNING nova.compute.manager [req-829e001b-2d24-43e2-a691-44c371ad1ada req-c88deba6-a8c7-487a-89ec-91153df8958d service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received unexpected event network-vif-plugged-e34b5242-a01d-49e3-9aba-aaf54bf279df for instance with vm_state building and task_state spawning. [ 835.655858] env[61907]: DEBUG oslo_concurrency.lockutils [None req-065e11b3-1d73-4aad-b570-500579e15ce7 tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "5b20f363-dfad-4c2b-b757-4da4e5bd0b99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.421s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.668265] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Successfully updated port: e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.767754] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243868, 'name': Rename_Task, 'duration_secs': 0.139243} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.768233] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 835.772501] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0da467c8-70ae-4a00-8d3e-7ad4cf3b8dea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.774774] env[61907]: DEBUG nova.network.neutron [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Updated VIF entry in instance network info cache for port 1773082b-e928-4ca3-ab1c-515606cb7e12. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.775172] env[61907]: DEBUG nova.network.neutron [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Updating instance_info_cache with network_info: [{"id": "1773082b-e928-4ca3-ab1c-515606cb7e12", "address": "fa:16:3e:0b:b9:a2", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1773082b-e9", "ovs_interfaceid": "1773082b-e928-4ca3-ab1c-515606cb7e12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.786700] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243869, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.787688] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 835.787688] env[61907]: value = "task-1243870" [ 835.787688] env[61907]: _type = "Task" [ 835.787688] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.800070] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243870, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.842877] env[61907]: DEBUG nova.network.neutron [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updated VIF entry in instance network info cache for port f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 835.842877] env[61907]: DEBUG nova.network.neutron [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.893186] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.999842] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d186b1a-a438-446b-abf6-743de08324ec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.008231] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07290f28-63d1-46ea-acfc-6c36e0a7ba84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.039101] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14e62e7f-51e0-4093-a53d-e3e3d44360c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.047316] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9d50ba-95e2-42dd-a527-832a44561489 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.063949] env[61907]: DEBUG nova.compute.provider_tree [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.173395] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.173395] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.173395] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.238450] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.279551] env[61907]: DEBUG oslo_concurrency.lockutils [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] Releasing lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.279934] env[61907]: DEBUG nova.compute.manager [req-07034a50-ce3a-4724-a1e4-9a643ba7be4e req-092d74b2-a059-4a6a-93e1-3276d572e485 service nova] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Received event network-vif-deleted-0f0768ac-bf89-4dc8-ba37-3e163bd551b1 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 836.280290] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243869, 'name': CreateVM_Task, 'duration_secs': 0.576767} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.280444] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.281119] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.281293] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.281621] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.281887] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da4fb188-c0fd-4a11-bad6-0212c280bc7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.287265] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 836.287265] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e52c3c-d9c4-6235-443a-bdf5b02b53f6" [ 836.287265] env[61907]: _type = "Task" [ 836.287265] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.300146] env[61907]: DEBUG oslo_vmware.api [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243870, 'name': PowerOnVM_Task, 'duration_secs': 0.475909} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.303582] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 836.303831] env[61907]: INFO nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Took 8.81 seconds to spawn the instance on the hypervisor. [ 836.303983] env[61907]: DEBUG nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 836.304297] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e52c3c-d9c4-6235-443a-bdf5b02b53f6, 'name': SearchDatastore_Task, 'duration_secs': 0.010781} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.304978] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0b17dd-6954-475c-bf0e-159df0533809 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.307394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.307595] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.307811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.308021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.308164] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.308399] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-edab03d6-1188-48d6-9605-dbad413f621b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.317476] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.317476] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 836.318301] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ece5150-8045-4d2f-ab51-099f4ede5475 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.323021] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 836.323021] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b07ebd-3798-42e7-9197-19ac5f648e38" [ 836.323021] env[61907]: _type = "Task" [ 836.323021] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.330546] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b07ebd-3798-42e7-9197-19ac5f648e38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.345349] env[61907]: DEBUG oslo_concurrency.lockutils [req-efa5ec84-ccff-4832-a063-9bdcd6848d98 req-e728137d-c0de-4914-a9f6-2166ef427078 service nova] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.567539] env[61907]: DEBUG nova.scheduler.client.report [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 836.684366] env[61907]: DEBUG nova.compute.manager [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 836.684519] env[61907]: DEBUG nova.compute.manager [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing instance network info cache due to event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 836.684720] env[61907]: DEBUG oslo_concurrency.lockutils [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.707754] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 836.828644] env[61907]: INFO nova.compute.manager [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Took 32.36 seconds to build instance. [ 836.841039] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b07ebd-3798-42e7-9197-19ac5f648e38, 'name': SearchDatastore_Task, 'duration_secs': 0.011698} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.841868] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57efd7e8-a7ac-4952-b8c9-71efc67bcfae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.848315] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 836.848315] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f25b57-b7d7-0412-8326-7124b54d000e" [ 836.848315] env[61907]: _type = "Task" [ 836.848315] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.856485] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f25b57-b7d7-0412-8326-7124b54d000e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.878307] env[61907]: DEBUG nova.network.neutron [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.073905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.074831] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.836s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.076845] env[61907]: INFO nova.compute.claims [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.099547] env[61907]: INFO nova.scheduler.client.report [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Deleted allocations for instance 08647e48-9e24-4d7c-a9fd-4066317a7250 [ 837.331360] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d2ebb589-1138-4ac8-b69e-28ea4e5d7a16 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.460s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.359457] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f25b57-b7d7-0412-8326-7124b54d000e, 'name': SearchDatastore_Task, 'duration_secs': 0.008886} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.359729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.360021] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/23526967-cd8f-4581-b9c5-1c270d385163.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 837.360387] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-28f63426-5e5d-4ad0-9895-3105ac8f2b8d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.368464] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 837.368464] env[61907]: value = "task-1243871" [ 837.368464] env[61907]: _type = "Task" [ 837.368464] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.376488] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.381490] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.381841] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Instance network_info: |[{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 837.382270] env[61907]: DEBUG oslo_concurrency.lockutils [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.382507] env[61907]: DEBUG nova.network.neutron [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.383812] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:45:7b:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e59b364d-b7f6-499d-b7dc-82b8a819aa12', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e34b5242-a01d-49e3-9aba-aaf54bf279df', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.392756] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Creating folder: Project (8e704030091a4d7985062c2a39d4f1b5). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.393887] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd281663-690d-4df7-80a3-26f35133bad3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.404398] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Created folder: Project (8e704030091a4d7985062c2a39d4f1b5) in parent group-v268168. [ 837.404647] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Creating folder: Instances. Parent ref: group-v268248. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.404948] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f8555ee-721b-4a33-860e-c5486042a7bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.413848] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Created folder: Instances in parent group-v268248. [ 837.414158] env[61907]: DEBUG oslo.service.loopingcall [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.414411] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.414653] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-804f4be7-9157-4580-8c91-63d781726484 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.433062] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.433062] env[61907]: value = "task-1243874" [ 837.433062] env[61907]: _type = "Task" [ 837.433062] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.440467] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243874, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.608420] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b540bf26-27a9-4a5e-ace9-30453e59405d tempest-MultipleCreateTestJSON-2124629329 tempest-MultipleCreateTestJSON-2124629329-project-member] Lock "08647e48-9e24-4d7c-a9fd-4066317a7250" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.271s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.833975] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 837.881755] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.449983} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.882043] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/23526967-cd8f-4581-b9c5-1c270d385163.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 837.882366] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.882579] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a95b6428-9d3f-4211-979f-0447dc322289 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.889434] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 837.889434] env[61907]: value = "task-1243875" [ 837.889434] env[61907]: _type = "Task" [ 837.889434] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.899095] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243875, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.948332] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243874, 'name': CreateVM_Task, 'duration_secs': 0.413774} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.948597] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.949649] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.949843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.950198] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.950497] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2030eec-824d-432d-9fbe-16e686c026c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.956496] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 837.956496] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5260cbd1-8185-b7d5-4ac3-ee6f4fcfe830" [ 837.956496] env[61907]: _type = "Task" [ 837.956496] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.965422] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5260cbd1-8185-b7d5-4ac3-ee6f4fcfe830, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.129327] env[61907]: DEBUG nova.network.neutron [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updated VIF entry in instance network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.129327] env[61907]: DEBUG nova.network.neutron [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.357565] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.373220] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371c01c5-4737-4a02-ba31-2a72904b43f9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.382065] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba52d460-6eea-4f61-98cd-63e058be78f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.418797] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b347cd7f-fc6f-4fde-9554-8e213823797a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.425477] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243875, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07375} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.427462] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.428241] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ebf8bf8-28fb-448b-ac63-68b63cfc002d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.431325] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd77090-4545-47e3-b4dd-d3bdb8b8bdd8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.444016] env[61907]: DEBUG nova.compute.provider_tree [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.464238] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/23526967-cd8f-4581-b9c5-1c270d385163.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.465532] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37f60295-d75a-4d46-a272-1f6b634ed8c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.489097] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5260cbd1-8185-b7d5-4ac3-ee6f4fcfe830, 'name': SearchDatastore_Task, 'duration_secs': 0.010262} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.490334] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.490571] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.490812] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.490966] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.491174] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.491480] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 838.491480] env[61907]: value = "task-1243876" [ 838.491480] env[61907]: _type = "Task" [ 838.491480] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.491665] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bffc182d-b45c-4abd-94b4-b488344959e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.501520] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243876, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.503310] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.503498] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.504225] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-110bda6f-b57f-446f-9a27-739d999f324d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.509013] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 838.509013] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f99227-b813-fe40-2080-0557342ca43f" [ 838.509013] env[61907]: _type = "Task" [ 838.509013] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.516388] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f99227-b813-fe40-2080-0557342ca43f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.630843] env[61907]: DEBUG oslo_concurrency.lockutils [req-dadd1420-f136-4d78-b69d-1d40934209eb req-e142e675-1710-4ab7-9381-4b13a18fab52 service nova] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.965903] env[61907]: DEBUG nova.scheduler.client.report [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 839.006240] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.025025] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f99227-b813-fe40-2080-0557342ca43f, 'name': SearchDatastore_Task, 'duration_secs': 0.008453} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.025974] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-367c50ec-edc7-476e-8a02-e0f2f3c5cbd9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.031854] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 839.031854] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f6102-5126-057e-333e-4b32edc93a19" [ 839.031854] env[61907]: _type = "Task" [ 839.031854] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.040939] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f6102-5126-057e-333e-4b32edc93a19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.470828] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.396s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.471419] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 839.474998] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.120s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.476935] env[61907]: INFO nova.compute.claims [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 839.510433] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243876, 'name': ReconfigVM_Task, 'duration_secs': 0.794635} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.510433] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/23526967-cd8f-4581-b9c5-1c270d385163.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.510433] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-79f21d97-02f0-43b7-9419-8636ab0f2a4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.517983] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 839.517983] env[61907]: value = "task-1243877" [ 839.517983] env[61907]: _type = "Task" [ 839.517983] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.536592] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243877, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.545376] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f6102-5126-057e-333e-4b32edc93a19, 'name': SearchDatastore_Task, 'duration_secs': 0.009681} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.545700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.545997] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] da490943-511b-4776-8f16-4f51c3b055c4/da490943-511b-4776-8f16-4f51c3b055c4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.546452] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca46806a-8c2b-4423-9f8b-be86f52ee200 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.554965] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 839.554965] env[61907]: value = "task-1243878" [ 839.554965] env[61907]: _type = "Task" [ 839.554965] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.567718] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.982967] env[61907]: DEBUG nova.compute.utils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.985132] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 839.985132] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.028998] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243877, 'name': Rename_Task, 'duration_secs': 0.168714} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.028998] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.029368] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f023a36d-4473-4e02-8c60-bd76a9a9d4a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.035582] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 840.035582] env[61907]: value = "task-1243879" [ 840.035582] env[61907]: _type = "Task" [ 840.035582] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.042907] env[61907]: DEBUG nova.policy [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '575a48c6fb4d44638edf42db3763e78a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a85baf8058442b39a07436a046e1ebe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.049301] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.064356] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.454371} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.064608] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] da490943-511b-4776-8f16-4f51c3b055c4/da490943-511b-4776-8f16-4f51c3b055c4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.064836] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.065113] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d73fea7-d3c3-486a-834a-b3ebad5c265a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.071425] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 840.071425] env[61907]: value = "task-1243880" [ 840.071425] env[61907]: _type = "Task" [ 840.071425] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.080293] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243880, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.390355] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Successfully created port: 5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.488401] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 840.548609] env[61907]: DEBUG oslo_vmware.api [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243879, 'name': PowerOnVM_Task, 'duration_secs': 0.464543} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.549037] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 840.553175] env[61907]: INFO nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Took 7.89 seconds to spawn the instance on the hypervisor. [ 840.553417] env[61907]: DEBUG nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 840.554522] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f754c0-896e-4c94-9aee-78e859003fa3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.583918] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243880, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.208518} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.586464] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.587886] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944fbc17-db94-4a20-9ad4-b9816026647e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.612922] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] da490943-511b-4776-8f16-4f51c3b055c4/da490943-511b-4776-8f16-4f51c3b055c4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.617029] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42afe429-e3bf-4168-965a-1651ee162e82 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.639149] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 840.639149] env[61907]: value = "task-1243881" [ 840.639149] env[61907]: _type = "Task" [ 840.639149] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.649766] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.860870] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35aa65c3-19c2-455b-83fa-9b8595df801b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.869328] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e151105f-0218-46a2-bc5d-8d2d990e78c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.903332] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28f7da2-6e5c-47fa-9baa-e0602311bc99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.911991] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39b1e45-2474-482a-a952-38304cfe853f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.928942] env[61907]: DEBUG nova.compute.provider_tree [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.077787] env[61907]: INFO nova.compute.manager [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Took 33.17 seconds to build instance. [ 841.150357] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243881, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.434025] env[61907]: DEBUG nova.scheduler.client.report [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 841.500744] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 841.529819] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.530106] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.530278] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.530484] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.530712] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.530883] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.531188] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.531376] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.531569] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.531741] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.531922] env[61907]: DEBUG nova.virt.hardware [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.532905] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c19359-4b2c-4dbf-a323-34fd4c7b5a69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.540964] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aef86532-7562-4721-8efb-547949f356eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.580415] env[61907]: DEBUG oslo_concurrency.lockutils [None req-868b8bcf-2058-492a-ac90-4126bbfce0a4 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.939s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.650797] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243881, 'name': ReconfigVM_Task, 'duration_secs': 0.907323} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.651319] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfigured VM instance instance-00000042 to attach disk [datastore1] da490943-511b-4776-8f16-4f51c3b055c4/da490943-511b-4776-8f16-4f51c3b055c4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.651864] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed152826-97d8-400a-a16f-4796023a687c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.658278] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 841.658278] env[61907]: value = "task-1243882" [ 841.658278] env[61907]: _type = "Task" [ 841.658278] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.666733] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243882, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.927101] env[61907]: DEBUG nova.compute.manager [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Received event network-vif-plugged-5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 841.927101] env[61907]: DEBUG oslo_concurrency.lockutils [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.927101] env[61907]: DEBUG oslo_concurrency.lockutils [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.927101] env[61907]: DEBUG oslo_concurrency.lockutils [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.928523] env[61907]: DEBUG nova.compute.manager [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] No waiting events found dispatching network-vif-plugged-5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 841.928893] env[61907]: WARNING nova.compute.manager [req-858d6a18-a329-48e1-b9a9-8064c7d20104 req-b0abd2f2-285d-4b8f-be64-4fb54f853fa0 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Received unexpected event network-vif-plugged-5290ad3b-915d-49f5-b5e8-57438dd6ca7b for instance with vm_state building and task_state spawning. [ 841.937432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.938473] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 841.942303] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 24.454s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.999936] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Successfully updated port: 5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.085051] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 842.171253] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243882, 'name': Rename_Task, 'duration_secs': 0.135813} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.171253] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.171253] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fd292b4-723a-4009-9089-ff1f4fac0036 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.178460] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 842.178460] env[61907]: value = "task-1243883" [ 842.178460] env[61907]: _type = "Task" [ 842.178460] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.187221] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243883, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.262812] env[61907]: INFO nova.compute.manager [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Rescuing [ 842.262812] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.262968] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.263151] env[61907]: DEBUG nova.network.neutron [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.457716] env[61907]: DEBUG nova.compute.utils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 842.465345] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 842.465548] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 842.509145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.511179] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.511679] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.526574] env[61907]: DEBUG nova.policy [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '84750bb751724a2f8e5ba3107f73defd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e6a77496c4f4c7ea060d39ad1504c53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 842.614863] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.690098] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243883, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.843773] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Successfully created port: 9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 842.962888] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 842.994759] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 865d386d-c5f0-439b-b50b-b613993ec9b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.994927] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ce864af6-aef6-4044-be64-8440ba175438 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.995079] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 52404147-0375-4f9c-9e1e-4a52001df7cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.995215] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f157c0db-054b-4d99-803c-11aa40cf0670 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.995340] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.995460] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e5d4890e-be1b-4b28-9721-89ef678c1182 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.995576] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5c908f9c-7efb-4588-b6bf-70babb173767 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.995694] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e1f8f88b-7179-4e92-83cd-a1027ff9b0a2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 842.997047] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.997047] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.997047] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23526967-cd8f-4581-b9c5-1c270d385163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.997047] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance da490943-511b-4776-8f16-4f51c3b055c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.997212] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 842.997212] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ddcbf498-fd27-41f2-bbcc-23a8f9827823 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 843.109188] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.189922] env[61907]: DEBUG oslo_vmware.api [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1243883, 'name': PowerOnVM_Task, 'duration_secs': 0.902391} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.190406] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.190690] env[61907]: INFO nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Took 8.10 seconds to spawn the instance on the hypervisor. [ 843.190939] env[61907]: DEBUG nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 843.191915] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728204ab-551e-477b-817e-bb59e96950c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.505406] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 41c680fb-5450-43f3-9acb-5218e57a7da9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 843.561216] env[61907]: DEBUG nova.network.neutron [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Updating instance_info_cache with network_info: [{"id": "1773082b-e928-4ca3-ab1c-515606cb7e12", "address": "fa:16:3e:0b:b9:a2", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1773082b-e9", "ovs_interfaceid": "1773082b-e928-4ca3-ab1c-515606cb7e12", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.628061] env[61907]: DEBUG nova.network.neutron [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.715822] env[61907]: INFO nova.compute.manager [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Took 33.59 seconds to build instance. [ 843.979654] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 844.009386] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.013024] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.013024] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.013024] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.013024] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.013024] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.013245] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.013245] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.013245] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.013245] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.013245] env[61907]: DEBUG nova.virt.hardware [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.013383] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f028beaf-4185-40d7-a730-abe91209f9d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.016543] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b611785b-a752-4ebd-9ebd-cd610c1b86c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.022940] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45969335-2e93-4fe2-8975-c6ab8c35975e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.064147] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-23526967-cd8f-4581-b9c5-1c270d385163" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.091976] env[61907]: DEBUG nova.compute.manager [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Received event network-changed-5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 844.092385] env[61907]: DEBUG nova.compute.manager [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Refreshing instance network info cache due to event network-changed-5290ad3b-915d-49f5-b5e8-57438dd6ca7b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 844.092472] env[61907]: DEBUG oslo_concurrency.lockutils [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] Acquiring lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.130604] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.130946] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Instance network_info: |[{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 844.131285] env[61907]: DEBUG oslo_concurrency.lockutils [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] Acquired lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.131474] env[61907]: DEBUG nova.network.neutron [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Refreshing network info cache for port 5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.132758] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:b6:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53955f0e-c162-4cef-8bd5-335b369c36b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5290ad3b-915d-49f5-b5e8-57438dd6ca7b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.140441] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Creating folder: Project (5a85baf8058442b39a07436a046e1ebe). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.143314] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b29116f5-aac0-4a32-862b-2f5fe43b1b3c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.155690] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Created folder: Project (5a85baf8058442b39a07436a046e1ebe) in parent group-v268168. [ 844.159020] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Creating folder: Instances. Parent ref: group-v268251. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.159020] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ce2bb34-6958-428d-ae3e-1e8caf405057 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.166581] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Created folder: Instances in parent group-v268251. [ 844.166839] env[61907]: DEBUG oslo.service.loopingcall [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.167050] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.167316] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b202238-6cbb-4233-a732-53ea87c0f9fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.188555] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.188555] env[61907]: value = "task-1243886" [ 844.188555] env[61907]: _type = "Task" [ 844.188555] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.199045] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243886, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.221485] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fae821fa-a829-48a5-880f-47f09dd3ad80 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.267s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.408742] env[61907]: DEBUG nova.network.neutron [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updated VIF entry in instance network info cache for port 5290ad3b-915d-49f5-b5e8-57438dd6ca7b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 844.409234] env[61907]: DEBUG nova.network.neutron [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.518135] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2c1212a7-df13-49b7-b3b3-24fb59b6789d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 844.698028] env[61907]: DEBUG nova.compute.manager [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 844.698231] env[61907]: DEBUG nova.compute.manager [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing instance network info cache due to event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 844.698439] env[61907]: DEBUG oslo_concurrency.lockutils [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.698585] env[61907]: DEBUG oslo_concurrency.lockutils [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.698743] env[61907]: DEBUG nova.network.neutron [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.705467] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243886, 'name': CreateVM_Task, 'duration_secs': 0.363005} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.705467] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 844.706204] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.706697] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.707090] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.708634] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2bf6f1e-5568-4f70-a31a-e6f384c3b537 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.714090] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 844.714090] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521d8675-9edf-9089-dfac-0363d35db5a2" [ 844.714090] env[61907]: _type = "Task" [ 844.714090] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.727230] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 844.728764] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521d8675-9edf-9089-dfac-0363d35db5a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.917267] env[61907]: DEBUG oslo_concurrency.lockutils [req-3e335610-bb41-49b5-910a-c2e47c2f528d req-987d2267-d6a1-40b5-9f4f-334b23f59b3d service nova] Releasing lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.021165] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.058654] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "5c357582-0b55-4de7-968e-4daa3522dda4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.058898] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.080885] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Successfully updated port: 9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.096475] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.097466] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a5fadf06-66da-4e9b-8a08-4ecb8b894054 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.106216] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 845.106216] env[61907]: value = "task-1243887" [ 845.106216] env[61907]: _type = "Task" [ 845.106216] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.114919] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243887, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.225992] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521d8675-9edf-9089-dfac-0363d35db5a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010245} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.226470] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.226734] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.226985] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.227153] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.227341] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.227616] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5936d71-1ebb-4d5a-9591-f2a9c2d83180 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.239019] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.239019] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.239689] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-925ed876-f1f6-4c81-b268-3e6189e787b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.246024] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 845.246024] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5223bb04-2053-a501-4235-9f100d1f30ec" [ 845.246024] env[61907]: _type = "Task" [ 845.246024] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.246983] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.255358] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5223bb04-2053-a501-4235-9f100d1f30ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.495324] env[61907]: DEBUG nova.network.neutron [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updated VIF entry in instance network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.495324] env[61907]: DEBUG nova.network.neutron [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.528061] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance aec01156-431b-447e-88ef-cb8d254c6341 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 845.586303] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.586969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.586969] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 845.617568] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243887, 'name': PowerOffVM_Task, 'duration_secs': 0.350187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.617852] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.618768] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2404c2b2-392a-4352-abfd-ea291b59a1ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.638637] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d1bde7-0dfe-4285-af47-a22602e0e4a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.675215] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.675536] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7097216f-07f1-43b5-9bd6-a725ffb962ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.682055] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 845.682055] env[61907]: value = "task-1243888" [ 845.682055] env[61907]: _type = "Task" [ 845.682055] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.689635] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243888, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.757414] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5223bb04-2053-a501-4235-9f100d1f30ec, 'name': SearchDatastore_Task, 'duration_secs': 0.013397} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.758325] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0045a2a3-69a2-4adb-a78b-1f3a92afbcea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.764339] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 845.764339] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523e4b3b-19cd-5c20-d7f9-8fb3dd2daf75" [ 845.764339] env[61907]: _type = "Task" [ 845.764339] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.771554] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523e4b3b-19cd-5c20-d7f9-8fb3dd2daf75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.997150] env[61907]: DEBUG oslo_concurrency.lockutils [req-db02ef14-8010-4cec-b50b-36afa88c06aa req-e7da6790-3aa7-4e4a-8c29-7223a3d9a977 service nova] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.033390] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance d32315e1-68b9-49ee-9393-8b2f9cbaf645 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.124560] env[61907]: DEBUG nova.compute.manager [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Received event network-vif-plugged-9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 846.124851] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Acquiring lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.125241] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.125426] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.125613] env[61907]: DEBUG nova.compute.manager [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] No waiting events found dispatching network-vif-plugged-9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 846.125788] env[61907]: WARNING nova.compute.manager [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Received unexpected event network-vif-plugged-9f84bd0f-4a0a-43d1-b913-e6667f455b55 for instance with vm_state building and task_state spawning. [ 846.125954] env[61907]: DEBUG nova.compute.manager [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Received event network-changed-9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 846.126131] env[61907]: DEBUG nova.compute.manager [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Refreshing instance network info cache due to event network-changed-9f84bd0f-4a0a-43d1-b913-e6667f455b55. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 846.126305] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Acquiring lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.139966] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.193032] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 846.193329] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 846.193581] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.193738] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.193924] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 846.194196] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8cf759cd-5bf8-4dae-8a12-757253da7302 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.204311] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 846.204528] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 846.205293] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08ecf8db-7e91-4a1c-9544-6a728d03e6ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.213153] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 846.213153] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e1eb95-20c7-9456-d616-4bc1b37ba27a" [ 846.213153] env[61907]: _type = "Task" [ 846.213153] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.220276] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e1eb95-20c7-9456-d616-4bc1b37ba27a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.273617] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523e4b3b-19cd-5c20-d7f9-8fb3dd2daf75, 'name': SearchDatastore_Task, 'duration_secs': 0.009712} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.273961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.274196] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 4c70ef23-fd26-4e90-b227-13aa050ff46d/4c70ef23-fd26-4e90-b227-13aa050ff46d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.274452] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0d9a379b-2306-4d3d-9d53-320c5ad1c808 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.279189] env[61907]: DEBUG nova.network.neutron [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updating instance_info_cache with network_info: [{"id": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "address": "fa:16:3e:14:87:fa", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f84bd0f-4a", "ovs_interfaceid": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.281408] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 846.281408] env[61907]: value = "task-1243889" [ 846.281408] env[61907]: _type = "Task" [ 846.281408] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.289013] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243889, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.537019] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e52648ac-e068-4a63-baa0-a1c34df52197 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 846.724141] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e1eb95-20c7-9456-d616-4bc1b37ba27a, 'name': SearchDatastore_Task, 'duration_secs': 0.010553} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.725045] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14931243-8db2-46ea-95bc-764540d30a71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.731285] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 846.731285] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528fe3a1-1f6e-90d4-0214-e4e27964f336" [ 846.731285] env[61907]: _type = "Task" [ 846.731285] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.738874] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528fe3a1-1f6e-90d4-0214-e4e27964f336, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.782696] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.783057] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Instance network_info: |[{"id": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "address": "fa:16:3e:14:87:fa", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f84bd0f-4a", "ovs_interfaceid": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 846.783529] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Acquired lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.783729] env[61907]: DEBUG nova.network.neutron [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Refreshing network info cache for port 9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 846.785392] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:14:87:fa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f84bd0f-4a0a-43d1-b913-e6667f455b55', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 846.792947] env[61907]: DEBUG oslo.service.loopingcall [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.796887] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 846.797170] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-337d7997-6809-43c6-b65d-26cbbe83c240 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.818719] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243889, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417521} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.819973] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 4c70ef23-fd26-4e90-b227-13aa050ff46d/4c70ef23-fd26-4e90-b227-13aa050ff46d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.820242] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.820474] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 846.820474] env[61907]: value = "task-1243890" [ 846.820474] env[61907]: _type = "Task" [ 846.820474] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.820662] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c23d7a89-d0d2-4b29-b532-7811705d8dcc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.831115] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243890, 'name': CreateVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.832432] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 846.832432] env[61907]: value = "task-1243891" [ 846.832432] env[61907]: _type = "Task" [ 846.832432] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.840574] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.041220] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 37ff6c54-6b79-4a9c-bc16-29d974185d3e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 847.041220] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (91bca385-a423-4ca4-9da0-aeb4615e22d3): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 847.041220] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 847.041220] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 847.244065] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528fe3a1-1f6e-90d4-0214-e4e27964f336, 'name': SearchDatastore_Task, 'duration_secs': 0.009387} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.244366] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.244664] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 847.244945] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fac2711b-a33c-40a0-8079-79d227aa6c1c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.251457] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 847.251457] env[61907]: value = "task-1243892" [ 847.251457] env[61907]: _type = "Task" [ 847.251457] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.259323] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243892, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.335677] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243890, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.341917] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662adaea-d462-4bd5-912f-3d82215c9111 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.348135] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064399} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.348813] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.349704] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2b60d91-76cb-48f4-a44f-9e04ace1dd2e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.355288] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ba2987-d87f-4d18-9807-8d603eca5444 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.377284] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 4c70ef23-fd26-4e90-b227-13aa050ff46d/4c70ef23-fd26-4e90-b227-13aa050ff46d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 847.380278] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1de2af1d-a60c-43e1-83f0-1563f0ce149b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.420492] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ba9ca9-5219-41b1-ae4c-6e49076f0a67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.429188] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8eb08a-963d-4ec9-892a-2917e6d3a50e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.433117] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 847.433117] env[61907]: value = "task-1243893" [ 847.433117] env[61907]: _type = "Task" [ 847.433117] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.446574] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.453504] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243893, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.569225] env[61907]: DEBUG nova.network.neutron [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updated VIF entry in instance network info cache for port 9f84bd0f-4a0a-43d1-b913-e6667f455b55. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 847.569692] env[61907]: DEBUG nova.network.neutron [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updating instance_info_cache with network_info: [{"id": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "address": "fa:16:3e:14:87:fa", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f84bd0f-4a", "ovs_interfaceid": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.761519] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243892, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.834459] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243890, 'name': CreateVM_Task, 'duration_secs': 0.650295} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.834832] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 847.835684] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.835892] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.836252] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 847.836542] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05046342-4ffe-423b-854d-301b33e1a1fe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.841903] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 847.841903] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52223ec0-956f-0261-3a64-57cd242e933c" [ 847.841903] env[61907]: _type = "Task" [ 847.841903] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.850835] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52223ec0-956f-0261-3a64-57cd242e933c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.943232] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243893, 'name': ReconfigVM_Task, 'duration_secs': 0.478704} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.943563] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 4c70ef23-fd26-4e90-b227-13aa050ff46d/4c70ef23-fd26-4e90-b227-13aa050ff46d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.944215] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e4e7690-32fd-4128-b249-276ef17b2bc4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.951836] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 847.956339] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 847.956339] env[61907]: value = "task-1243894" [ 847.956339] env[61907]: _type = "Task" [ 847.956339] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.968383] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243894, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.072936] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4b35127-c8e6-4265-8873-0cb465dfecc9 req-edf359be-c7fe-4c83-a6b7-abf6a7212b7d service nova] Releasing lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.261434] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243892, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.516562} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.261710] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 848.262528] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c65172cf-41cd-4344-aa02-535f154b60d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.286533] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.286780] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0e8096d-27ab-4317-bb60-c4bf1de8e50d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.304741] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 848.304741] env[61907]: value = "task-1243895" [ 848.304741] env[61907]: _type = "Task" [ 848.304741] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.315227] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.352274] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52223ec0-956f-0261-3a64-57cd242e933c, 'name': SearchDatastore_Task, 'duration_secs': 0.010193} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.352644] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.352883] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 848.353159] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.353348] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.353553] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 848.354541] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a98f20ff-a226-4b54-a702-27e4dc084e58 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.361820] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 848.362014] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 848.362755] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-003a1b26-f928-4017-b9b7-48f57f32c958 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.367324] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 848.367324] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e51b1f-0aea-cc35-003e-ecec8b1a6bd1" [ 848.367324] env[61907]: _type = "Task" [ 848.367324] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.374409] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e51b1f-0aea-cc35-003e-ecec8b1a6bd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.458688] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 848.459036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.517s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.459559] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.621s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.461150] env[61907]: INFO nova.compute.claims [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.473594] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243894, 'name': Rename_Task, 'duration_secs': 0.162856} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.473594] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.473739] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16daef7a-1b20-4d34-8927-f1d23c43e5ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.480364] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 848.480364] env[61907]: value = "task-1243896" [ 848.480364] env[61907]: _type = "Task" [ 848.480364] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.488261] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.815525] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.881518] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e51b1f-0aea-cc35-003e-ecec8b1a6bd1, 'name': SearchDatastore_Task, 'duration_secs': 0.008243} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.883070] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcf70543-5925-4c59-87f0-372213c05a43 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.890667] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 848.890667] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524ef9f5-1f17-5026-40a9-cdf8c159a3f6" [ 848.890667] env[61907]: _type = "Task" [ 848.890667] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.899708] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524ef9f5-1f17-5026-40a9-cdf8c159a3f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.934567] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.934567] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 848.934567] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 848.934742] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 848.990424] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243896, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.315332] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243895, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.376278] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.376827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.377332] env[61907]: DEBUG nova.objects.instance [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'flavor' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.408077] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524ef9f5-1f17-5026-40a9-cdf8c159a3f6, 'name': SearchDatastore_Task, 'duration_secs': 0.009875} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.408506] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.408921] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/ddcbf498-fd27-41f2-bbcc-23a8f9827823.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 849.409316] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a454fb32-096b-41ee-95a7-95ebfdb26ac9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.417993] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 849.417993] env[61907]: value = "task-1243897" [ 849.417993] env[61907]: _type = "Task" [ 849.417993] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.430618] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243897, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.440586] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 849.440770] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 849.440934] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 849.441433] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.441433] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.441433] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 849.441645] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.495621] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243896, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.820466] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243895, 'name': ReconfigVM_Task, 'duration_secs': 1.196331} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.821047] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Reconfigured VM instance instance-00000041 to attach disk [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.822360] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9e10c4f-7852-4022-8b5f-daad264c5f5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.854616] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d70aaefa-5830-4b65-9f69-a591071af509 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.875184] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 849.875184] env[61907]: value = "task-1243898" [ 849.875184] env[61907]: _type = "Task" [ 849.875184] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.883295] env[61907]: DEBUG nova.objects.instance [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'pci_requests' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.887972] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6ab4eb-7782-41ab-be93-10dedf6d49cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.898725] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f9349c-1827-41b1-bc8c-696aecef115c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.903117] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243898, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.960281] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1e378b-4f7e-480a-911d-57fcae818e45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.976824] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243897, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539136} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.978853] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c6f8a93-a197-49c5-93b8-9976af60d125 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.986723] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/ddcbf498-fd27-41f2-bbcc-23a8f9827823.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.987124] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.988438] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-275fe72e-1937-42da-bf29-2d56615dc1b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.012293] env[61907]: DEBUG nova.compute.provider_tree [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.021243] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 850.021243] env[61907]: value = "task-1243899" [ 850.021243] env[61907]: _type = "Task" [ 850.021243] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.021671] env[61907]: DEBUG oslo_vmware.api [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1243896, 'name': PowerOnVM_Task, 'duration_secs': 1.12428} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.022567] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 850.022950] env[61907]: INFO nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Took 8.52 seconds to spawn the instance on the hypervisor. [ 850.023281] env[61907]: DEBUG nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 850.029719] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a52db30-c086-45bf-9463-f9bbe586c2e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.043033] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243899, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.385809] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243898, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.389498] env[61907]: DEBUG nova.objects.base [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Object Instance<52404147-0375-4f9c-9e1e-4a52001df7cb> lazy-loaded attributes: flavor,pci_requests {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 850.389709] env[61907]: DEBUG nova.network.neutron [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.481988] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8fcdfb4d-8b89-4a1f-acf4-ed21bb89e164 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.105s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.524115] env[61907]: DEBUG nova.scheduler.client.report [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 850.541314] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243899, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122747} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.542411] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 850.543617] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72065fa-e83a-46c8-9800-dfa5c75a2477 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.585504] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/ddcbf498-fd27-41f2-bbcc-23a8f9827823.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 850.588546] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2e16b2c-a9bc-4826-b4b4-b38de397043d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.612281] env[61907]: INFO nova.compute.manager [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Took 37.41 seconds to build instance. [ 850.622184] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 850.622184] env[61907]: value = "task-1243900" [ 850.622184] env[61907]: _type = "Task" [ 850.622184] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.637587] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243900, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.886796] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243898, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.037591] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.038154] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 851.041980] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.560s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.041980] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.043673] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 31.561s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.045216] env[61907]: INFO nova.compute.claims [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.077818] env[61907]: INFO nova.scheduler.client.report [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Deleted allocations for instance f157c0db-054b-4d99-803c-11aa40cf0670 [ 851.114784] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c98912b-b89f-4b61-9a5c-b2ce4c1c3f4a tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 80.969s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.135101] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243900, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.214905] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap40a79714-91", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.387995] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243898, 'name': ReconfigVM_Task, 'duration_secs': 1.201873} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.389073] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.389073] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cacdacee-4e56-4531-8e98-e1aebb015999 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.396627] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 851.396627] env[61907]: value = "task-1243901" [ 851.396627] env[61907]: _type = "Task" [ 851.396627] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.407066] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.552532] env[61907]: DEBUG nova.compute.utils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 851.554099] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 851.554285] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 851.586342] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fd7a5742-2c6f-464c-808e-a29cbc313b3f tempest-InstanceActionsV221TestJSON-1063708298 tempest-InstanceActionsV221TestJSON-1063708298-project-member] Lock "f157c0db-054b-4d99-803c-11aa40cf0670" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 36.086s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.600501] env[61907]: DEBUG nova.policy [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '779d389d7771450f9b5adc352a41f407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87439b1925b644888b9ed1a52ddff936', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 851.617458] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 851.634836] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243900, 'name': ReconfigVM_Task, 'duration_secs': 0.722821} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.635243] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfigured VM instance instance-00000044 to attach disk [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/ddcbf498-fd27-41f2-bbcc-23a8f9827823.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 851.635976] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e905923-6ac6-4bb7-ae9b-e42088f88221 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.644214] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 851.644214] env[61907]: value = "task-1243902" [ 851.644214] env[61907]: _type = "Task" [ 851.644214] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.654552] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243902, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.717631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.717950] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 851.718095] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.718263] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.718422] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.718571] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.718720] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 851.718852] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 851.888798] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Successfully created port: 4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.908399] env[61907]: DEBUG oslo_vmware.api [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243901, 'name': PowerOnVM_Task, 'duration_secs': 0.384026} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.908687] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 851.912146] env[61907]: DEBUG nova.compute.manager [None req-7cf7a12f-211e-4bea-861b-df8e5dfabbb3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 851.913489] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0875a0a-c3c4-4411-a888-91eb2db475c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.058114] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 852.109615] env[61907]: DEBUG nova.compute.manager [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Received event network-changed-5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 852.109837] env[61907]: DEBUG nova.compute.manager [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Refreshing instance network info cache due to event network-changed-5290ad3b-915d-49f5-b5e8-57438dd6ca7b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 852.110070] env[61907]: DEBUG oslo_concurrency.lockutils [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] Acquiring lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.110464] env[61907]: DEBUG oslo_concurrency.lockutils [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] Acquired lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.110764] env[61907]: DEBUG nova.network.neutron [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Refreshing network info cache for port 5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 852.144444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.159749] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243902, 'name': Rename_Task, 'duration_secs': 0.166776} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.160053] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 852.160409] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-467a3179-7f52-4b2b-94a0-8cee2f682c46 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.171703] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 852.171703] env[61907]: value = "task-1243903" [ 852.171703] env[61907]: _type = "Task" [ 852.171703] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.187113] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.432736] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c62972-84de-421a-9922-5441a12123df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.442902] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635b022e-4c24-4f9f-b6e2-9321b31344cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.481592] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30d2316b-ed88-49f4-a303-73feac9e05e1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.490860] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4940618f-a311-4433-bbae-e884d7a8103e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.509339] env[61907]: DEBUG nova.compute.provider_tree [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.598029] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.598029] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.598029] env[61907]: DEBUG nova.objects.instance [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'flavor' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.682271] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243903, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.013656] env[61907]: DEBUG nova.scheduler.client.report [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 853.068023] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 853.093345] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 853.093619] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 853.093785] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 853.093980] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 853.094285] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 853.094460] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 853.094725] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 853.094898] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 853.095093] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 853.095293] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 853.095507] env[61907]: DEBUG nova.virt.hardware [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 853.096427] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc21871a-53cc-45b8-bfa9-83494b9bcf99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.108330] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e1f297-3c46-4cb3-97a6-aa3cd0f4b504 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.182849] env[61907]: DEBUG oslo_vmware.api [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243903, 'name': PowerOnVM_Task, 'duration_secs': 0.875799} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.183157] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 853.183442] env[61907]: INFO nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Took 9.20 seconds to spawn the instance on the hypervisor. [ 853.183638] env[61907]: DEBUG nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 853.184428] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc7f99c-7623-4cf9-9329-c6115afe304d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.310597] env[61907]: DEBUG nova.network.neutron [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updated VIF entry in instance network info cache for port 5290ad3b-915d-49f5-b5e8-57438dd6ca7b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 853.310987] env[61907]: DEBUG nova.network.neutron [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.430594] env[61907]: DEBUG nova.objects.instance [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'pci_requests' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.513733] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 853.523199] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.523734] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 853.530023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 30.122s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.530023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.530023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.311s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.530332] env[61907]: INFO nova.compute.claims [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 853.555550] env[61907]: INFO nova.scheduler.client.report [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Deleted allocations for instance 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111 [ 853.701941] env[61907]: INFO nova.compute.manager [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Took 37.38 seconds to build instance. [ 853.814079] env[61907]: DEBUG oslo_concurrency.lockutils [req-e0204a15-5bf4-4c15-aba7-46280dced82d req-fef8d0d3-60e4-4da1-bb85-7939fe641ca8 service nova] Releasing lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.936196] env[61907]: DEBUG nova.objects.base [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Object Instance<52404147-0375-4f9c-9e1e-4a52001df7cb> lazy-loaded attributes: flavor,pci_requests {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 853.936196] env[61907]: DEBUG nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.024944] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.026399] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.026399] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 854.026399] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 854.035099] env[61907]: DEBUG nova.compute.utils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.038369] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 854.038677] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 854.045022] env[61907]: DEBUG nova.policy [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.069708] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c631d2c8-934c-4101-9fe1-3d64be65d602 tempest-ServerMetadataNegativeTestJSON-1698318303 tempest-ServerMetadataNegativeTestJSON-1698318303-project-member] Lock "7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.202s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.153510] env[61907]: DEBUG nova.policy [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '779d389d7771450f9b5adc352a41f407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87439b1925b644888b9ed1a52ddff936', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.204459] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031b1f76-bcc1-4d83-b43c-b4bb81aa38ab tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.867s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.426147] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Successfully updated port: 4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.488213] env[61907]: DEBUG nova.compute.manager [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Received event network-vif-plugged-4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 854.488481] env[61907]: DEBUG oslo_concurrency.lockutils [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] Acquiring lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.488733] env[61907]: DEBUG oslo_concurrency.lockutils [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.488945] env[61907]: DEBUG oslo_concurrency.lockutils [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.489212] env[61907]: DEBUG nova.compute.manager [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] No waiting events found dispatching network-vif-plugged-4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 854.489468] env[61907]: WARNING nova.compute.manager [req-4b1e20e9-f8bf-4a4d-8f70-a192d40a8336 req-15333d9c-03f0-4b96-a496-cbe42353baff service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Received unexpected event network-vif-plugged-4d440707-544e-46dc-bf63-b4910c232a7a for instance with vm_state building and task_state spawning. [ 854.528192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.549213] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 854.683146] env[61907]: DEBUG nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Successfully created port: 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.762120] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Successfully created port: b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 854.907392] env[61907]: DEBUG nova.compute.manager [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Received event network-changed-4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 854.907596] env[61907]: DEBUG nova.compute.manager [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Refreshing instance network info cache due to event network-changed-4d440707-544e-46dc-bf63-b4910c232a7a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 854.907812] env[61907]: DEBUG oslo_concurrency.lockutils [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] Acquiring lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.907952] env[61907]: DEBUG oslo_concurrency.lockutils [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] Acquired lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.908129] env[61907]: DEBUG nova.network.neutron [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Refreshing network info cache for port 4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.929574] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.930297] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db4cd08-b3b8-4cf9-b6f4-2e1de793a5d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.941445] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b436262d-5860-4f9a-8d11-555532b2e082 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.981884] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccfa417-c7a4-4942-9891-26757cca968a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.992394] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82816aa-2e29-47c5-9513-23cb1986c6e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.007263] env[61907]: DEBUG nova.compute.provider_tree [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.265541] env[61907]: INFO nova.compute.manager [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Rescuing [ 855.265851] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.266025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.266243] env[61907]: DEBUG nova.network.neutron [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.466427] env[61907]: DEBUG nova.network.neutron [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.512620] env[61907]: DEBUG nova.scheduler.client.report [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 855.561134] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 855.589915] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.592589] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.592589] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.592589] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.592589] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.592589] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.592850] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.592850] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.592850] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.592850] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.592850] env[61907]: DEBUG nova.virt.hardware [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.593421] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bab5a05-0eb4-44e5-a20d-f9f3888b828a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.603564] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f95b75-4611-4557-aa93-1a0145e8e913 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.674011] env[61907]: DEBUG nova.network.neutron [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.016789] env[61907]: DEBUG nova.network.neutron [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updating instance_info_cache with network_info: [{"id": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "address": "fa:16:3e:14:87:fa", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f84bd0f-4a", "ovs_interfaceid": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.019022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.021071] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 856.024036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.843s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.025443] env[61907]: INFO nova.compute.claims [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.178287] env[61907]: DEBUG oslo_concurrency.lockutils [req-77135534-6f7f-4ba3-8348-6bc0d93415b0 req-8126e587-9aeb-4bdb-aece-ad4eaf8b5957 service nova] Releasing lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.178701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.178855] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 856.528297] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.532273] env[61907]: DEBUG nova.compute.utils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 856.533794] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 856.534090] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 856.639847] env[61907]: DEBUG nova.policy [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '779d389d7771450f9b5adc352a41f407', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87439b1925b644888b9ed1a52ddff936', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 856.714131] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.868368] env[61907]: DEBUG nova.network.neutron [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Updating instance_info_cache with network_info: [{"id": "4d440707-544e-46dc-bf63-b4910c232a7a", "address": "fa:16:3e:15:8c:a3", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d440707-54", "ovs_interfaceid": "4d440707-544e-46dc-bf63-b4910c232a7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.950116] env[61907]: INFO nova.compute.manager [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Rescuing [ 856.950415] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.950582] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.950755] env[61907]: DEBUG nova.network.neutron [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 857.038725] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 857.349255] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54e08d0-ec6c-4e54-8dcb-43198dc48488 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.362125] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6e2e823-fd1f-45d0-8acc-dc208bc535ac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.370977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.371390] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance network_info: |[{"id": "4d440707-544e-46dc-bf63-b4910c232a7a", "address": "fa:16:3e:15:8c:a3", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d440707-54", "ovs_interfaceid": "4d440707-544e-46dc-bf63-b4910c232a7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 857.402786] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:8c:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '822050c7-1845-485d-b87e-73778d21c33c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4d440707-544e-46dc-bf63-b4910c232a7a', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 857.410791] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Creating folder: Project (87439b1925b644888b9ed1a52ddff936). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.411374] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fe684d19-3cff-4de8-b108-a0be83d8d9b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.413802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedc9aa7-4fd9-4951-8c49-610dc6649d14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.423026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b24be15-dd4c-4990-a615-045dc6d21d11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.428870] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Created folder: Project (87439b1925b644888b9ed1a52ddff936) in parent group-v268168. [ 857.429091] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Creating folder: Instances. Parent ref: group-v268255. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 857.430217] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e149ecc2-f395-4a7a-8438-07e0f719e002 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.441486] env[61907]: DEBUG nova.compute.provider_tree [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.454218] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Created folder: Instances in parent group-v268255. [ 857.454573] env[61907]: DEBUG oslo.service.loopingcall [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 857.455557] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 857.455837] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-659093b9-74ce-408d-a9ca-da7e1e4ad229 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.473923] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Successfully updated port: b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.487300] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 857.487300] env[61907]: value = "task-1243906" [ 857.487300] env[61907]: _type = "Task" [ 857.487300] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.497442] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243906, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.569354] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 857.570012] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e1f636a-9336-440a-86e9-f2fcef83a150 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.578214] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 857.578214] env[61907]: value = "task-1243907" [ 857.578214] env[61907]: _type = "Task" [ 857.578214] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.589801] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.603769] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Successfully created port: e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 857.908494] env[61907]: DEBUG nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Successfully updated port: 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.947045] env[61907]: DEBUG nova.scheduler.client.report [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 857.976200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.976337] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.976493] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.000624] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243906, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.048617] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 858.053234] env[61907]: DEBUG nova.network.neutron [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.087373] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 858.087637] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 858.087797] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 858.087980] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 858.088145] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 858.088295] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 858.089082] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 858.089082] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 858.089082] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 858.089082] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 858.089262] env[61907]: DEBUG nova.virt.hardware [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 858.090255] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e193a88d-12d6-46fd-b0c6-5ab01a3f4e16 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.096043] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243907, 'name': PowerOffVM_Task, 'duration_secs': 0.270777} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.096459] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.097238] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d6d3c8e-4322-4daf-afed-e2a8826e7bfc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.103253] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f762faa-a51e-4816-9b93-4cce29a44837 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.125475] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a704d2dc-1ef1-4d43-b38a-0e956ef5abe4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.168633] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.168967] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bbe1d433-2752-4ccf-b19d-032a9445e2f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.176980] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 858.176980] env[61907]: value = "task-1243908" [ 858.176980] env[61907]: _type = "Task" [ 858.176980] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.185617] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243908, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.187583] env[61907]: DEBUG nova.compute.manager [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Received event network-vif-plugged-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 858.187732] env[61907]: DEBUG oslo_concurrency.lockutils [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] Acquiring lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.187943] env[61907]: DEBUG oslo_concurrency.lockutils [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] Lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.188131] env[61907]: DEBUG oslo_concurrency.lockutils [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] Lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.188302] env[61907]: DEBUG nova.compute.manager [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] No waiting events found dispatching network-vif-plugged-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 858.188472] env[61907]: WARNING nova.compute.manager [req-618e0527-134b-42e0-91db-5512ea940689 req-56df371a-4231-4660-898d-58bbd88bb2ee service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Received unexpected event network-vif-plugged-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb for instance with vm_state building and task_state spawning. [ 858.410677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.410846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.411027] env[61907]: DEBUG nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 858.453293] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.453899] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 858.456735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.381s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.456851] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.459338] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 27.343s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.459552] env[61907]: DEBUG nova.objects.instance [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 858.482593] env[61907]: INFO nova.scheduler.client.report [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted allocations for instance ce864af6-aef6-4044-be64-8440ba175438 [ 858.503236] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243906, 'name': CreateVM_Task, 'duration_secs': 0.575727} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.503672] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 858.504464] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.504729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.505111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 858.505912] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a6f914ac-4ca8-4ae7-91a4-ff9fa77da54f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.511582] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 858.516364] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 858.516364] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b0adaa-1227-7fab-7d88-f4acb8cc4805" [ 858.516364] env[61907]: _type = "Task" [ 858.516364] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.530175] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b0adaa-1227-7fab-7d88-f4acb8cc4805, 'name': SearchDatastore_Task, 'duration_secs': 0.012991} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.530582] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.530834] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.531086] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.531241] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.531442] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.531700] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7b5901a-1839-4af4-bfa3-362ba5e8fa07 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.540849] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.541050] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.541778] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-392c8a6a-5d46-47be-9e70-f1c13d1d6ba6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.549356] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 858.549356] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522118aa-1ffe-0b09-8e37-85d730784a28" [ 858.549356] env[61907]: _type = "Task" [ 858.549356] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.555279] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.564513] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522118aa-1ffe-0b09-8e37-85d730784a28, 'name': SearchDatastore_Task, 'duration_secs': 0.010549} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.565871] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa836bde-6a93-4289-a639-9b6b37a4ac9a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.573025] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 858.573025] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520f4c16-ed6c-9eeb-62a6-0ad307f2e850" [ 858.573025] env[61907]: _type = "Task" [ 858.573025] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.587850] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520f4c16-ed6c-9eeb-62a6-0ad307f2e850, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.591569] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.591862] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 41c680fb-5450-43f3-9acb-5218e57a7da9/41c680fb-5450-43f3-9acb-5218e57a7da9.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.594475] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56459b0a-8989-4a25-af3f-f8e9e49d12bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.605874] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 858.605874] env[61907]: value = "task-1243909" [ 858.605874] env[61907]: _type = "Task" [ 858.605874] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.617192] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.687257] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 858.687518] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 858.687813] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.687971] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.688063] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.688345] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d0b33f5-a6d5-4e30-a9b4-9c2f314fbef6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.696481] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.696677] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.697448] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4fc3502-5004-4395-93a2-3303fb10f2d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.705429] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 858.705429] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527e7198-2e9a-034b-9bea-1fff30ab82db" [ 858.705429] env[61907]: _type = "Task" [ 858.705429] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.714119] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527e7198-2e9a-034b-9bea-1fff30ab82db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.752310] env[61907]: DEBUG nova.network.neutron [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Updating instance_info_cache with network_info: [{"id": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "address": "fa:16:3e:ec:bb:2e", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2a3b1ca-0e", "ovs_interfaceid": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.967234] env[61907]: DEBUG nova.compute.utils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.972561] env[61907]: WARNING nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] 9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b already exists in list: networks containing: ['9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b']. ignoring it [ 858.975516] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 858.975971] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.995835] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ba1113bc-e8fe-43c5-9d8d-c8a0c138c4e0 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "ce864af6-aef6-4044-be64-8440ba175438" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.544s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.084187] env[61907]: DEBUG nova.policy [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41fc76e7ed9a46a09f96f11b28556ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b57c1194e0f14ae498d99d302da346c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 859.118960] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481793} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.119267] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 41c680fb-5450-43f3-9acb-5218e57a7da9/41c680fb-5450-43f3-9acb-5218e57a7da9.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.119487] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.119754] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3eda77ea-2159-4615-9026-bddf5c8994eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.127775] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 859.127775] env[61907]: value = "task-1243910" [ 859.127775] env[61907]: _type = "Task" [ 859.127775] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.137131] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.218306] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527e7198-2e9a-034b-9bea-1fff30ab82db, 'name': SearchDatastore_Task, 'duration_secs': 0.009187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.219390] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f44dd1e8-d697-4e6d-bb6b-5f0675b308ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.225973] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 859.225973] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b4f922-d4fe-93ae-d822-87beae47045b" [ 859.225973] env[61907]: _type = "Task" [ 859.225973] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.235710] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b4f922-d4fe-93ae-d822-87beae47045b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.254831] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.255188] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Instance network_info: |[{"id": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "address": "fa:16:3e:ec:bb:2e", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2a3b1ca-0e", "ovs_interfaceid": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 859.256220] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ec:bb:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '822050c7-1845-485d-b87e-73778d21c33c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 859.264631] env[61907]: DEBUG oslo.service.loopingcall [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 859.264865] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 859.265127] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a54dfec7-57a7-44bf-9b00-8a6e553dc8cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.287577] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 859.287577] env[61907]: value = "task-1243911" [ 859.287577] env[61907]: _type = "Task" [ 859.287577] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.298607] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243911, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.473208] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 859.480282] env[61907]: DEBUG oslo_concurrency.lockutils [None req-031a80c2-61cf-48f3-b301-5be694532227 tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.481345] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.713s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.481550] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.485843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.239s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.487895] env[61907]: INFO nova.compute.claims [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.601372] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 859.601372] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24f385ce-300a-4d9d-8746-70eeb4c5939f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.607878] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 859.607878] env[61907]: value = "task-1243912" [ 859.607878] env[61907]: _type = "Task" [ 859.607878] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.623520] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.637445] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07697} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.637960] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.638896] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a1c0fe3-3964-4f83-9a1c-864716855893 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.663793] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] 41c680fb-5450-43f3-9acb-5218e57a7da9/41c680fb-5450-43f3-9acb-5218e57a7da9.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.664278] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fc1cf73b-9f6b-4c11-b3c3-a88bc78152ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.691023] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 859.691023] env[61907]: value = "task-1243913" [ 859.691023] env[61907]: _type = "Task" [ 859.691023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.700065] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243913, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.740250] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b4f922-d4fe-93ae-d822-87beae47045b, 'name': SearchDatastore_Task, 'duration_secs': 0.010393} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.740861] env[61907]: DEBUG oslo_concurrency.lockutils [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.741235] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 859.741665] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea1bef89-f992-427e-b274-833930eb91c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.753052] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 859.753052] env[61907]: value = "task-1243914" [ 859.753052] env[61907]: _type = "Task" [ 859.753052] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.764427] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.797871] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243911, 'name': CreateVM_Task, 'duration_secs': 0.3455} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.798294] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 859.799182] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.799490] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.799914] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.800317] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57e6ffab-6002-4694-8130-4b0636aae29a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.805701] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 859.805701] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528cc581-3fd8-03ea-dcec-3aa6e5ccae9f" [ 859.805701] env[61907]: _type = "Task" [ 859.805701] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.815152] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528cc581-3fd8-03ea-dcec-3aa6e5ccae9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.858762] env[61907]: DEBUG nova.network.neutron [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "address": "fa:16:3e:9f:37:bf", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c78d5e6-cc", "ovs_interfaceid": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.947570] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Successfully created port: 6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.002394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0efd8079-1b56-4428-a85a-dcba9a4f7eef tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 46.446s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.005204] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 23.767s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.005575] env[61907]: INFO nova.compute.manager [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Unshelving [ 860.057258] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Successfully updated port: e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.126022] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243912, 'name': PowerOffVM_Task, 'duration_secs': 0.275878} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.126022] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 860.126022] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c79597-67f4-4459-acbe-aafa5992dafa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.145701] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d406b5d-51a0-44c5-9bb9-b1be261b280b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.188540] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 860.188892] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a13cb354-7528-443e-b114-88fbf44675a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.202378] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243913, 'name': ReconfigVM_Task, 'duration_secs': 0.303649} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.204119] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Reconfigured VM instance instance-00000045 to attach disk [datastore1] 41c680fb-5450-43f3-9acb-5218e57a7da9/41c680fb-5450-43f3-9acb-5218e57a7da9.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.204897] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 860.204897] env[61907]: value = "task-1243915" [ 860.204897] env[61907]: _type = "Task" [ 860.204897] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.205123] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fd46ca8e-f676-4848-b4cf-f6c0ed8e5ca4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.219065] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 860.219312] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.219566] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.219723] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.219905] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.220312] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 860.220312] env[61907]: value = "task-1243916" [ 860.220312] env[61907]: _type = "Task" [ 860.220312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.220534] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d716a63-e679-4db8-b553-7a4785a87c3e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.231760] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243916, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.233925] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.234191] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.234910] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90b834b1-6599-493f-9a44-8d3ba867e020 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.240186] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 860.240186] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dcec5e-9a38-e841-8f32-c1417d05ba30" [ 860.240186] env[61907]: _type = "Task" [ 860.240186] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.249126] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dcec5e-9a38-e841-8f32-c1417d05ba30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.260929] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243914, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484564} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.261723] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 860.262560] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bee092f-b3d1-4fd8-b778-7ed6eee17d3e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.288135] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.288476] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b18925ec-7b69-4c61-9c69-bb1fcd6ec2a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.319997] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528cc581-3fd8-03ea-dcec-3aa6e5ccae9f, 'name': SearchDatastore_Task, 'duration_secs': 0.021564} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.319997] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.319997] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 860.320345] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.320677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.320935] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 860.321470] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 860.321470] env[61907]: value = "task-1243917" [ 860.321470] env[61907]: _type = "Task" [ 860.321470] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.321788] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7024a2d8-1743-4746-bf1d-bbb1bd7e95d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.334841] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.336814] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 860.336814] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 860.337436] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b99589b-d131-41c3-969d-65aaafe1c83d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.343840] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 860.343840] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52228490-0c46-5a39-bc33-24382e44cc96" [ 860.343840] env[61907]: _type = "Task" [ 860.343840] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.354512] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52228490-0c46-5a39-bc33-24382e44cc96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.360233] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.360904] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.361079] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.361869] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bee2cbf9-400e-4dc7-8b65-50073da93786 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.379124] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.379411] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.379577] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.379763] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.379958] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.380204] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.380423] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.380609] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.380787] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.380996] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.381259] env[61907]: DEBUG nova.virt.hardware [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.387438] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfiguring VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 860.387818] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92487077-2b47-46b5-afc0-1c9afb27aec8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.406404] env[61907]: DEBUG oslo_vmware.api [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 860.406404] env[61907]: value = "task-1243918" [ 860.406404] env[61907]: _type = "Task" [ 860.406404] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.417046] env[61907]: DEBUG oslo_vmware.api [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243918, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.485279] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 860.514045] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 860.514377] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 860.514377] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 860.514477] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 860.515325] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 860.515325] env[61907]: DEBUG nova.virt.hardware [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 860.515752] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e89132-b214-43dc-8205-84fea6af8977 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.527924] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88d0b2a-2763-433c-9988-0a1995e7f54c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.560777] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.560926] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.561087] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.616557] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Received event network-changed-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 860.616717] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Refreshing instance network info cache due to event network-changed-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 860.616807] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquiring lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.616944] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquired lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.617203] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Refreshing network info cache for port b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 860.733524] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243916, 'name': Rename_Task, 'duration_secs': 0.170796} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.733897] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.736448] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4137801a-ec6c-4042-b12c-a75eb334ad54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.745445] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 860.745445] env[61907]: value = "task-1243919" [ 860.745445] env[61907]: _type = "Task" [ 860.745445] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.756306] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52dcec5e-9a38-e841-8f32-c1417d05ba30, 'name': SearchDatastore_Task, 'duration_secs': 0.015423} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.757138] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e312d7f-893a-43a6-983a-74ff4ad4f3ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.763233] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243919, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.770472] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 860.770472] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52313c13-41f0-363b-17d9-9bc26f52a719" [ 860.770472] env[61907]: _type = "Task" [ 860.770472] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.783945] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52313c13-41f0-363b-17d9-9bc26f52a719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.811966] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8c58d5a-c5d1-41d5-aefd-5e698e4c95a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.820346] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd931e7-c76e-4567-a9cd-91ca578648e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.856984] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eee18d59-6d63-4485-9370-f0457a203d14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.862802] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243917, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.868924] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52228490-0c46-5a39-bc33-24382e44cc96, 'name': SearchDatastore_Task, 'duration_secs': 0.010317} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.871626] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84aad2ff-6960-4e60-b21c-3417b937139a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.874955] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aed22fc6-9e2e-420d-8db2-4c56f11e3501 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.882603] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 860.882603] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52697e42-e25a-811e-e2cd-c35edd6ef721" [ 860.882603] env[61907]: _type = "Task" [ 860.882603] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.891144] env[61907]: DEBUG nova.compute.provider_tree [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 860.901160] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52697e42-e25a-811e-e2cd-c35edd6ef721, 'name': SearchDatastore_Task, 'duration_secs': 0.00976} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.901998] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.902281] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f028beaf-4185-40d7-a730-abe91209f9d8/f028beaf-4185-40d7-a730-abe91209f9d8.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 860.902584] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ccaa6b72-bc46-4a61-8b3f-abf3f80c9b87 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.911565] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 860.911565] env[61907]: value = "task-1243920" [ 860.911565] env[61907]: _type = "Task" [ 860.911565] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.918339] env[61907]: DEBUG oslo_vmware.api [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.924013] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243920, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.036383] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.109095] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.259888] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243919, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.285472] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52313c13-41f0-363b-17d9-9bc26f52a719, 'name': SearchDatastore_Task, 'duration_secs': 0.010927} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.285798] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.286100] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 861.286456] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e96be75c-cedc-427c-95f2-116342c8555a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.296242] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 861.296242] env[61907]: value = "task-1243921" [ 861.296242] env[61907]: _type = "Task" [ 861.296242] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.309410] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243921, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.337473] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243917, 'name': ReconfigVM_Task, 'duration_secs': 0.813232} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.337908] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfigured VM instance instance-00000044 to attach disk [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.339470] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64ca6939-cf80-4756-a12c-f3ce8e0324c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.372551] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9696a8e1-7876-4c36-9fc1-19ab6e11a2e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.396360] env[61907]: DEBUG nova.scheduler.client.report [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 861.399739] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 861.399739] env[61907]: value = "task-1243922" [ 861.399739] env[61907]: _type = "Task" [ 861.399739] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.411660] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243922, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.424963] env[61907]: DEBUG oslo_vmware.api [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.428733] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243920, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497988} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.429015] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f028beaf-4185-40d7-a730-abe91209f9d8/f028beaf-4185-40d7-a730-abe91209f9d8.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 861.429279] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.429553] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1607f98a-b048-4a71-bc68-e2bfed062f21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.437900] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 861.437900] env[61907]: value = "task-1243923" [ 861.437900] env[61907]: _type = "Task" [ 861.437900] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.449485] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.521715] env[61907]: DEBUG nova.network.neutron [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Updating instance_info_cache with network_info: [{"id": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "address": "fa:16:3e:4a:43:f1", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0833683-f8", "ovs_interfaceid": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.699792] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Updated VIF entry in instance network info cache for port b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 861.699917] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Updating instance_info_cache with network_info: [{"id": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "address": "fa:16:3e:ec:bb:2e", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb2a3b1ca-0e", "ovs_interfaceid": "b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.760428] env[61907]: DEBUG oslo_vmware.api [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243919, 'name': PowerOnVM_Task, 'duration_secs': 0.819733} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.760720] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.760988] env[61907]: INFO nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Took 8.69 seconds to spawn the instance on the hypervisor. [ 861.761244] env[61907]: DEBUG nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 861.762165] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9c050c-b44d-4b63-8b59-40d7bbbe3de6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.809941] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243921, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.901532] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.902109] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 861.905093] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 27.038s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.905304] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.912363] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.739s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.912617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.918895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.026s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.918895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.920328] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.563s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.926538] env[61907]: INFO nova.compute.claims [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 861.938112] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Successfully updated port: 6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 861.955729] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243922, 'name': ReconfigVM_Task, 'duration_secs': 0.302845} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.956915] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.957590] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-64dc6def-a931-47ba-89d2-a7cb8f8a1d44 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.967666] env[61907]: DEBUG oslo_vmware.api [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243918, 'name': ReconfigVM_Task, 'duration_secs': 1.291232} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.967666] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070685} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.968546] env[61907]: INFO nova.scheduler.client.report [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Deleted allocations for instance e1f8f88b-7179-4e92-83cd-a1027ff9b0a2 [ 861.970544] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.970670] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfigured VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 861.973540] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.977131] env[61907]: INFO nova.scheduler.client.report [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Deleted allocations for instance 5c908f9c-7efb-4588-b6bf-70babb173767 [ 861.981674] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f0dd6d-ff4d-4529-b6ee-672542774d3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.986375] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 861.986375] env[61907]: value = "task-1243924" [ 861.986375] env[61907]: _type = "Task" [ 861.986375] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.987267] env[61907]: INFO nova.scheduler.client.report [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Deleted allocations for instance e5d4890e-be1b-4b28-9721-89ef678c1182 [ 862.015566] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] f028beaf-4185-40d7-a730-abe91209f9d8/f028beaf-4185-40d7-a730-abe91209f9d8.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.016969] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-832d9b13-8899-476b-b5ea-17f87324e240 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.034963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.035269] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Instance network_info: |[{"id": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "address": "fa:16:3e:4a:43:f1", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0833683-f8", "ovs_interfaceid": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 862.035646] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243924, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.035946] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:43:f1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '822050c7-1845-485d-b87e-73778d21c33c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e0833683-f8a7-41f3-bca6-307d94fffbfd', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.044331] env[61907]: DEBUG oslo.service.loopingcall [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.046856] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.047121] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a41f7a5-da43-46bb-9be3-b6ba30d3154a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.068427] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 862.068427] env[61907]: value = "task-1243925" [ 862.068427] env[61907]: _type = "Task" [ 862.068427] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.075740] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.075740] env[61907]: value = "task-1243926" [ 862.075740] env[61907]: _type = "Task" [ 862.075740] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.083550] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.088542] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243926, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.203779] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Releasing lock "refresh_cache-f028beaf-4185-40d7-a730-abe91209f9d8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.204198] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-plugged-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 862.204457] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.204745] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.204973] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.205220] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] No waiting events found dispatching network-vif-plugged-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 862.205466] env[61907]: WARNING nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received unexpected event network-vif-plugged-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed for instance with vm_state active and task_state None. [ 862.205689] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-changed-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 862.205909] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing instance network info cache due to event network-changed-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 862.206190] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.206401] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.206630] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing network info cache for port 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.281900] env[61907]: INFO nova.compute.manager [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Took 43.47 seconds to build instance. [ 862.313368] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243921, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555134} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.313877] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 862.314890] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0365cca6-df02-4ea6-830d-26f36eb45d1b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.345416] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 862.346157] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47625875-8a6d-4763-858d-00bc4b123d3c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.367383] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 862.367383] env[61907]: value = "task-1243927" [ 862.367383] env[61907]: _type = "Task" [ 862.367383] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.377310] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243927, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.439185] env[61907]: DEBUG nova.compute.utils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.440749] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 862.441163] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.443182] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.443235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.443380] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.475440] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c6ff1d76-13d3-4595-aa96-1be4570df42f tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.880s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.490889] env[61907]: DEBUG oslo_concurrency.lockutils [None req-afab397d-3e35-4e33-82fd-58212378f321 tempest-VolumesAdminNegativeTest-903724568 tempest-VolumesAdminNegativeTest-903724568-project-member] Lock "e1f8f88b-7179-4e92-83cd-a1027ff9b0a2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.145s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.492559] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6a8837dc-c5b5-45b3-b290-a6c0b9a574da tempest-SecurityGroupsTestJSON-356670551 tempest-SecurityGroupsTestJSON-356670551-project-member] Lock "5c908f9c-7efb-4588-b6bf-70babb173767" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.368s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.505630] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243924, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.508997] env[61907]: DEBUG nova.policy [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'def3c7c3271940ffb8feae32e999ff00', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc568a1103574dd498a0e783b945fc78', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.516953] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b4150279-f749-44bf-a8db-0d41773493ee tempest-ServerShowV257Test-1558011625 tempest-ServerShowV257Test-1558011625-project-member] Lock "e5d4890e-be1b-4b28-9721-89ef678c1182" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 30.909s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.590753] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243925, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.595896] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243926, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.673107] env[61907]: DEBUG nova.compute.manager [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Received event network-vif-plugged-6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 862.673419] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.673672] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.673923] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.674047] env[61907]: DEBUG nova.compute.manager [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] No waiting events found dispatching network-vif-plugged-6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 862.674221] env[61907]: WARNING nova.compute.manager [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Received unexpected event network-vif-plugged-6007e3ea-4d8d-4523-b3da-93d26643424f for instance with vm_state building and task_state spawning. [ 862.674384] env[61907]: DEBUG nova.compute.manager [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Received event network-changed-6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 862.674537] env[61907]: DEBUG nova.compute.manager [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Refreshing instance network info cache due to event network-changed-6007e3ea-4d8d-4523-b3da-93d26643424f. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 862.674706] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Acquiring lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.784579] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b22569e9-8173-4aa8-8690-91d2f296c6c3 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.335s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.881405] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243927, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.945889] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 862.973221] env[61907]: DEBUG nova.scheduler.client.report [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 862.999362] env[61907]: DEBUG nova.scheduler.client.report [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 862.999684] env[61907]: DEBUG nova.compute.provider_tree [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.009016] env[61907]: DEBUG oslo_vmware.api [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243924, 'name': PowerOnVM_Task, 'duration_secs': 0.835797} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.009339] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 863.012131] env[61907]: DEBUG nova.compute.manager [None req-200f66e7-f819-4bcf-a6f4-30396e84930f tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 863.012982] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fce6e7-f044-4c7d-b32e-515a34e42140 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.016620] env[61907]: DEBUG nova.scheduler.client.report [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 863.038249] env[61907]: DEBUG nova.scheduler.client.report [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 863.046376] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.082045] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243925, 'name': ReconfigVM_Task, 'duration_secs': 0.735475} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.087950] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Reconfigured VM instance instance-00000046 to attach disk [datastore1] f028beaf-4185-40d7-a730-abe91209f9d8/f028beaf-4185-40d7-a730-abe91209f9d8.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.088855] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad30006d-6036-49d5-942a-0b1dd29966bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.097642] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243926, 'name': CreateVM_Task, 'duration_secs': 0.544732} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.098834] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.099233] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.099233] env[61907]: value = "task-1243928" [ 863.099233] env[61907]: _type = "Task" [ 863.099233] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.099952] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.100190] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.100592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.104189] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4fbd928-3aea-43f6-b032-4b5468298e26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.117970] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.117970] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4b77f-f8e1-47a4-ca22-3773e1ffbfc5" [ 863.117970] env[61907]: _type = "Task" [ 863.117970] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.123073] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243928, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.139672] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4b77f-f8e1-47a4-ca22-3773e1ffbfc5, 'name': SearchDatastore_Task, 'duration_secs': 0.01295} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.142811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.143267] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.143578] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.144428] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.144428] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.144735] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7808ae20-897e-4f8f-92d8-f310969e1f1a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.156950] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.157411] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.160137] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4382554e-f124-4ed9-a2b5-d031764d4c3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.167250] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.167250] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521a2188-8331-937e-d984-fe7a12247194" [ 863.167250] env[61907]: _type = "Task" [ 863.167250] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.180935] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521a2188-8331-937e-d984-fe7a12247194, 'name': SearchDatastore_Task, 'duration_secs': 0.010464} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.181886] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f26d89f-8c3f-4edf-ae2d-a300a6cb186a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.193951] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.193951] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528994a0-abc3-76b2-057e-31ca7fa42e47" [ 863.193951] env[61907]: _type = "Task" [ 863.193951] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.206231] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528994a0-abc3-76b2-057e-31ca7fa42e47, 'name': SearchDatastore_Task} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.206546] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.206854] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 2c1212a7-df13-49b7-b3b3-24fb59b6789d/2c1212a7-df13-49b7-b3b3-24fb59b6789d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 863.207173] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2167c8fb-19bb-49be-bc0d-e1fa89b37237 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.218380] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.218380] env[61907]: value = "task-1243929" [ 863.218380] env[61907]: _type = "Task" [ 863.218380] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.228694] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243929, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.357081] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ac0081-7f4e-4a47-8a8a-8353b1d4b434 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.369595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5709807-1181-4dcf-830f-2cd89b74d850 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.395720] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243927, 'name': ReconfigVM_Task, 'duration_secs': 0.630635} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.433909] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 863.435425] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334f124a-6e4d-4ff1-a21c-042ee35a6c99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.438757] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d95101-94b1-4715-b0a9-6a58a7f44490 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.480997] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cad71e51-38fd-4339-8ccd-e22ffbd3659b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.492364] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da8227d5-028e-4882-a3e5-990a7ffe0d74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.511042] env[61907]: DEBUG nova.compute.provider_tree [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.513919] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 863.513919] env[61907]: value = "task-1243930" [ 863.513919] env[61907]: _type = "Task" [ 863.513919] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.526536] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243930, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.589213] env[61907]: DEBUG nova.network.neutron [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.613880] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243928, 'name': Rename_Task, 'duration_secs': 0.299722} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.614454] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 863.615225] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5684f88b-6459-420f-812d-267046ca6df0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.626141] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.626141] env[61907]: value = "task-1243931" [ 863.626141] env[61907]: _type = "Task" [ 863.626141] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.637641] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.658971] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updated VIF entry in instance network info cache for port 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.659533] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "address": "fa:16:3e:9f:37:bf", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c78d5e6-cc", "ovs_interfaceid": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.710271] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Successfully created port: aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.730801] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243929, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502703} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.731054] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 2c1212a7-df13-49b7-b3b3-24fb59b6789d/2c1212a7-df13-49b7-b3b3-24fb59b6789d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 863.731282] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 863.731549] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09e1d00e-2138-464c-ae41-1060ff357219 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.738392] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 863.738392] env[61907]: value = "task-1243932" [ 863.738392] env[61907]: _type = "Task" [ 863.738392] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.747606] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243932, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.982870] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 864.020034] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.020305] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.020495] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.020709] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.020866] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.022347] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.022987] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.023259] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.023661] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.024183] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.024183] env[61907]: DEBUG nova.virt.hardware [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.027647] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f03516-5369-4386-9803-2c370b63f74e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.041978] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243930, 'name': ReconfigVM_Task, 'duration_secs': 0.211799} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.044610] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 864.044940] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3599a80-f0d0-4896-ba3d-1d4bf260acdd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.047962] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d3194d-5670-48bd-b6b6-e6a8e7a35447 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.058687] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 864.058687] env[61907]: value = "task-1243933" [ 864.058687] env[61907]: _type = "Task" [ 864.058687] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.067933] env[61907]: DEBUG nova.scheduler.client.report [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 864.068286] env[61907]: DEBUG nova.compute.provider_tree [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 87 to 88 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 864.068638] env[61907]: DEBUG nova.compute.provider_tree [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 864.082466] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243933, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.095021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.095021] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Instance network_info: |[{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 864.095021] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Acquired lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.095021] env[61907]: DEBUG nova.network.neutron [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Refreshing network info cache for port 6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.095021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:01:4f:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16c6ea68-9b0e-4ac0-a484-7a9a40533017', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6007e3ea-4d8d-4523-b3da-93d26643424f', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.106422] env[61907]: DEBUG oslo.service.loopingcall [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.106422] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.106630] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-03c23985-2f4b-4de5-a395-f6460f400097 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.132663] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.132663] env[61907]: value = "task-1243934" [ 864.132663] env[61907]: _type = "Task" [ 864.132663] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.136750] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243931, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.146595] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243934, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.162642] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.162919] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Received event network-vif-plugged-e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 864.163160] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquiring lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.163373] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.163676] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.163920] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] No waiting events found dispatching network-vif-plugged-e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 864.164133] env[61907]: WARNING nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Received unexpected event network-vif-plugged-e0833683-f8a7-41f3-bca6-307d94fffbfd for instance with vm_state building and task_state spawning. [ 864.164325] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Received event network-changed-e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 864.164492] env[61907]: DEBUG nova.compute.manager [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Refreshing instance network info cache due to event network-changed-e0833683-f8a7-41f3-bca6-307d94fffbfd. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 864.164694] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquiring lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.164844] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Acquired lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.165079] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Refreshing network info cache for port e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.249875] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243932, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068271} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.250210] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 864.251053] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a7523f-0164-452a-84c4-38ea6dfbbd8b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.279375] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 2c1212a7-df13-49b7-b3b3-24fb59b6789d/2c1212a7-df13-49b7-b3b3-24fb59b6789d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.279375] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eded2685-718e-4136-9c15-fac688e92db1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.298084] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 864.298084] env[61907]: value = "task-1243935" [ 864.298084] env[61907]: _type = "Task" [ 864.298084] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.309125] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243935, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.579090] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.658s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.579090] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 864.582502] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243933, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.582502] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.967s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.585719] env[61907]: INFO nova.compute.claims [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.639242] env[61907]: DEBUG oslo_vmware.api [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243931, 'name': PowerOnVM_Task, 'duration_secs': 0.729421} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.643639] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 864.644287] env[61907]: INFO nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Took 9.08 seconds to spawn the instance on the hypervisor. [ 864.644604] env[61907]: DEBUG nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 864.646082] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d8ce276-d99e-4dc8-b1f8-33e4644a5bc5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.666792] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243934, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.810915] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243935, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.852419] env[61907]: INFO nova.compute.manager [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Unrescuing [ 864.852419] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.852419] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquired lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.852419] env[61907]: DEBUG nova.network.neutron [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.041469] env[61907]: DEBUG nova.network.neutron [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updated VIF entry in instance network info cache for port 6007e3ea-4d8d-4523-b3da-93d26643424f. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.041469] env[61907]: DEBUG nova.network.neutron [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.083399] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Updated VIF entry in instance network info cache for port e0833683-f8a7-41f3-bca6-307d94fffbfd. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.084629] env[61907]: DEBUG nova.network.neutron [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Updating instance_info_cache with network_info: [{"id": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "address": "fa:16:3e:4a:43:f1", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape0833683-f8", "ovs_interfaceid": "e0833683-f8a7-41f3-bca6-307d94fffbfd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.087603] env[61907]: DEBUG nova.compute.utils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 865.093743] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 865.094082] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 865.098687] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243933, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.153232] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243934, 'name': CreateVM_Task, 'duration_secs': 0.553222} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.153764] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.154363] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.154844] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.155401] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.155854] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-524fc27f-8616-430c-bf78-df9d9be5cbee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.162309] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 865.162309] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3593-339c-c1e4-d7e8-8fbdd4d0b0ee" [ 865.162309] env[61907]: _type = "Task" [ 865.162309] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.173329] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3593-339c-c1e4-d7e8-8fbdd4d0b0ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.183418] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-b0ff98c3-b17f-497e-8ba2-0578ebbd2927" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.183836] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-b0ff98c3-b17f-497e-8ba2-0578ebbd2927" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.184279] env[61907]: DEBUG nova.objects.instance [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'flavor' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.187844] env[61907]: INFO nova.compute.manager [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Took 45.72 seconds to build instance. [ 865.191191] env[61907]: DEBUG nova.policy [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da98c262c7fb41dea5785c9b89f7fdb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2048329d91224454805d73742252fc4d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 865.309403] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243935, 'name': ReconfigVM_Task, 'duration_secs': 0.857993} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.309874] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 2c1212a7-df13-49b7-b3b3-24fb59b6789d/2c1212a7-df13-49b7-b3b3-24fb59b6789d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.310549] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6916d91b-655f-42a7-9f35-b02f161b3f7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.319394] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 865.319394] env[61907]: value = "task-1243936" [ 865.319394] env[61907]: _type = "Task" [ 865.319394] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.333363] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243936, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.501948] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Successfully updated port: aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.544484] env[61907]: DEBUG oslo_concurrency.lockutils [req-fa035254-da0d-45e7-bd66-22cb9842cfc1 req-2ab4bbd0-d0bd-456b-85ce-e641408e10d1 service nova] Releasing lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.583938] env[61907]: DEBUG oslo_vmware.api [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243933, 'name': PowerOnVM_Task, 'duration_secs': 1.095683} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.584546] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 865.587614] env[61907]: DEBUG nova.compute.manager [None req-6e233736-5717-4435-969b-c3d2f98f1871 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 865.588469] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc2713d2-0a2c-4d84-940e-b840db01d10a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.594144] env[61907]: DEBUG oslo_concurrency.lockutils [req-2345f659-26b3-4748-9a4b-58243a98e85d req-eb2d25ff-8f3e-4188-84a9-09d0c05f615c service nova] Releasing lock "refresh_cache-2c1212a7-df13-49b7-b3b3-24fb59b6789d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.602452] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 865.683230] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fc3593-339c-c1e4-d7e8-8fbdd4d0b0ee, 'name': SearchDatastore_Task, 'duration_secs': 0.011245} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.683230] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.683230] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.683230] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.683230] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.683230] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.683230] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f64b398-6034-4c92-80bb-91dabf305d4f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.692898] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0791859e-f45a-472b-a152-e67e979c2efb tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.540s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.702104] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.702104] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.705339] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-464f3d20-ab01-48d6-92bf-f7f4ecbd15bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.715027] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 865.715027] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525f25dc-81a8-3d48-d480-00eb854c0716" [ 865.715027] env[61907]: _type = "Task" [ 865.715027] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.725679] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525f25dc-81a8-3d48-d480-00eb854c0716, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.806013] env[61907]: DEBUG nova.compute.manager [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Received event network-vif-plugged-aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 865.806013] env[61907]: DEBUG oslo_concurrency.lockutils [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] Acquiring lock "aec01156-431b-447e-88ef-cb8d254c6341-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.806013] env[61907]: DEBUG oslo_concurrency.lockutils [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] Lock "aec01156-431b-447e-88ef-cb8d254c6341-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.806013] env[61907]: DEBUG oslo_concurrency.lockutils [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] Lock "aec01156-431b-447e-88ef-cb8d254c6341-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.806013] env[61907]: DEBUG nova.compute.manager [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] No waiting events found dispatching network-vif-plugged-aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 865.806013] env[61907]: WARNING nova.compute.manager [req-68557600-1762-469b-9ff5-51776bb473d4 req-90dd2b4a-67c3-402e-802c-beb94bcc910a service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Received unexpected event network-vif-plugged-aa81555f-9f10-410a-b9ad-12894784fd53 for instance with vm_state building and task_state spawning. [ 865.832898] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243936, 'name': Rename_Task, 'duration_secs': 0.291107} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.837724] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 865.838609] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b454142e-1f7a-4d60-879f-428510b8fc3c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.848231] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 865.848231] env[61907]: value = "task-1243937" [ 865.848231] env[61907]: _type = "Task" [ 865.848231] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.862255] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243937, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.900722] env[61907]: DEBUG nova.network.neutron [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updating instance_info_cache with network_info: [{"id": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "address": "fa:16:3e:14:87:fa", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f84bd0f-4a", "ovs_interfaceid": "9f84bd0f-4a0a-43d1-b913-e6667f455b55", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.931906] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df0c2ace-4be1-4540-aa6a-4a946aa857b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.942639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4eba76-a335-489d-99f3-bd760b2c869c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.978266] env[61907]: DEBUG nova.objects.instance [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'pci_requests' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.979758] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Successfully created port: 3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 865.988215] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757908c0-3e2e-4faf-8998-68ff89a1e885 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.997120] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9cee9c6-2a48-48ca-a78a-bb61b5be0f19 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.012167] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.012411] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquired lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.012482] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.014419] env[61907]: DEBUG nova.compute.provider_tree [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.086473] env[61907]: DEBUG nova.compute.manager [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Received event network-changed-aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 866.086736] env[61907]: DEBUG nova.compute.manager [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Refreshing instance network info cache due to event network-changed-aa81555f-9f10-410a-b9ad-12894784fd53. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 866.087368] env[61907]: DEBUG oslo_concurrency.lockutils [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] Acquiring lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.225871] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525f25dc-81a8-3d48-d480-00eb854c0716, 'name': SearchDatastore_Task, 'duration_secs': 0.010352} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.226789] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddeb3ad0-c5ef-4654-a476-617fa461f69a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.235045] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 866.235045] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529412b4-aaf6-4fdf-867a-34744d9b48ee" [ 866.235045] env[61907]: _type = "Task" [ 866.235045] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.245325] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529412b4-aaf6-4fdf-867a-34744d9b48ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.359880] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243937, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.403616] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Releasing lock "refresh_cache-ddcbf498-fd27-41f2-bbcc-23a8f9827823" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.404231] env[61907]: DEBUG nova.objects.instance [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lazy-loading 'flavor' on Instance uuid ddcbf498-fd27-41f2-bbcc-23a8f9827823 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.491627] env[61907]: DEBUG nova.objects.base [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Object Instance<52404147-0375-4f9c-9e1e-4a52001df7cb> lazy-loaded attributes: flavor,pci_requests {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 866.491885] env[61907]: DEBUG nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 866.518160] env[61907]: DEBUG nova.scheduler.client.report [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 866.614239] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 866.641804] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 866.642072] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 866.642243] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 866.642434] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 866.642586] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 866.642876] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 866.643113] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 866.643285] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 866.643459] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 866.643652] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 866.643884] env[61907]: DEBUG nova.virt.hardware [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 866.644819] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d23b61e8-a37c-4713-89a5-2b898ed0474a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.669165] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835a9331-55cd-406b-8c97-92dfa4a1d4a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.745895] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529412b4-aaf6-4fdf-867a-34744d9b48ee, 'name': SearchDatastore_Task, 'duration_secs': 0.011464} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.746200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.746466] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.746732] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f606fa46-afd6-4643-a370-0dd209a14bf8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.750930] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.757809] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 866.757809] env[61907]: value = "task-1243938" [ 866.757809] env[61907]: _type = "Task" [ 866.757809] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.770447] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243938, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.860512] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243937, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.901821] env[61907]: DEBUG nova.policy [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 866.913261] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d1547d3-9102-4d61-ac38-311f58a27445 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.941044] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 866.941237] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01eac6c1-ed18-4f67-97a3-fce31552ad02 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.950259] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 866.950259] env[61907]: value = "task-1243939" [ 866.950259] env[61907]: _type = "Task" [ 866.950259] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.960031] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243939, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.025972] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.026624] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 867.029623] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.783s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.031194] env[61907]: INFO nova.compute.claims [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.238635] env[61907]: DEBUG nova.network.neutron [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Updating instance_info_cache with network_info: [{"id": "aa81555f-9f10-410a-b9ad-12894784fd53", "address": "fa:16:3e:ee:16:e6", "network": {"id": "611325fa-e3e1-47f5-be0c-1495e46ed5b2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1884102614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc568a1103574dd498a0e783b945fc78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa81555f-9f", "ovs_interfaceid": "aa81555f-9f10-410a-b9ad-12894784fd53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.270259] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243938, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494368} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.270792] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.270933] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.271493] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-671201cf-d5d6-4e2a-bb2e-26961b7b036c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.281567] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 867.281567] env[61907]: value = "task-1243940" [ 867.281567] env[61907]: _type = "Task" [ 867.281567] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.289538] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243940, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.361569] env[61907]: DEBUG oslo_vmware.api [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243937, 'name': PowerOnVM_Task, 'duration_secs': 1.018832} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.361869] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.362291] env[61907]: INFO nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Took 9.31 seconds to spawn the instance on the hypervisor. [ 867.362528] env[61907]: DEBUG nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 867.363577] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fcffa8-21ae-4e43-8c35-e1d3e4ba0bfc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.469623] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243939, 'name': PowerOffVM_Task, 'duration_secs': 0.354865} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.469623] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 867.473057] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 867.473391] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9025cb69-95fe-449e-b5e8-32bac1c070b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.498077] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 867.498077] env[61907]: value = "task-1243941" [ 867.498077] env[61907]: _type = "Task" [ 867.498077] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.508150] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243941, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.538664] env[61907]: DEBUG nova.compute.utils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 867.540884] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 867.541339] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 867.665761] env[61907]: DEBUG nova.policy [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8682003dee14f25ba212739afd3d335', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6a88771a8f84196ae271dbd57590c15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 867.744406] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Releasing lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.744406] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Instance network_info: |[{"id": "aa81555f-9f10-410a-b9ad-12894784fd53", "address": "fa:16:3e:ee:16:e6", "network": {"id": "611325fa-e3e1-47f5-be0c-1495e46ed5b2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1884102614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc568a1103574dd498a0e783b945fc78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa81555f-9f", "ovs_interfaceid": "aa81555f-9f10-410a-b9ad-12894784fd53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 867.744630] env[61907]: DEBUG oslo_concurrency.lockutils [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] Acquired lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.744843] env[61907]: DEBUG nova.network.neutron [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Refreshing network info cache for port aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.748155] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:16:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '91c1da19-ab68-4127-bacd-accbaff19651', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa81555f-9f10-410a-b9ad-12894784fd53', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.758321] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Creating folder: Project (dc568a1103574dd498a0e783b945fc78). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.759777] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09e9d7e4-5cd3-4ac7-a844-4bdbd1ec496f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.772910] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Created folder: Project (dc568a1103574dd498a0e783b945fc78) in parent group-v268168. [ 867.773149] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Creating folder: Instances. Parent ref: group-v268261. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.773405] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8f43487-5d11-4cc1-abbe-35e70cd33625 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.786448] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Created folder: Instances in parent group-v268261. [ 867.788105] env[61907]: DEBUG oslo.service.loopingcall [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.788105] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.788105] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30446feb-4532-4746-944b-1fd1512504ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.806586] env[61907]: INFO nova.compute.manager [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Unrescuing [ 867.806842] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.806992] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.807169] env[61907]: DEBUG nova.network.neutron [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 867.808503] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243940, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078808} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.809259] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.810112] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8450040-b608-4a20-83a4-ef64b802af1f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.815125] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.815125] env[61907]: value = "task-1243944" [ 867.815125] env[61907]: _type = "Task" [ 867.815125] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.838020] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.839474] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2a79452-c09a-430b-933f-3c4d707bad35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.861018] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243944, 'name': CreateVM_Task} progress is 15%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.862885] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 867.862885] env[61907]: value = "task-1243945" [ 867.862885] env[61907]: _type = "Task" [ 867.862885] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.873595] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243945, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.885460] env[61907]: INFO nova.compute.manager [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Took 42.71 seconds to build instance. [ 868.008823] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243941, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.041714] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 868.239136] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Successfully updated port: 3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 868.330107] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243944, 'name': CreateVM_Task, 'duration_secs': 0.429837} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.330829] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 868.331441] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.331441] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.331729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 868.332033] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ccaac00-2ee5-4f9e-8186-b286883eb2a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.339194] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 868.339194] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c24d63-ee15-20f9-f03a-d1ca55b6477e" [ 868.339194] env[61907]: _type = "Task" [ 868.339194] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.348316] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bf2336-d98d-4887-9ccc-77ba297e2fce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.353973] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c24d63-ee15-20f9-f03a-d1ca55b6477e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.359645] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c047b0e8-0a9e-4d12-9aeb-36dbdc486ccd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.393839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c721f354-155f-41d0-b695-3d9ffeec00cd tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 68.129s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.396929] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Successfully created port: 9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 868.399460] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e14ea42a-9b5a-42aa-bf54-c275403f3b13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.410186] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.413773] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c70364a7-72ff-4092-9838-6f15b4a2213d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.429324] env[61907]: DEBUG nova.compute.provider_tree [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.432619] env[61907]: DEBUG nova.compute.manager [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Received event network-vif-plugged-3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 868.432619] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] Acquiring lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.432619] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.432619] env[61907]: DEBUG oslo_concurrency.lockutils [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.432839] env[61907]: DEBUG nova.compute.manager [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] No waiting events found dispatching network-vif-plugged-3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 868.432901] env[61907]: WARNING nova.compute.manager [req-8ce9ddc0-3b15-4ddd-83b3-cf798716a9c8 req-0402e563-56e3-4bd2-94b7-bbb666c7fc50 service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Received unexpected event network-vif-plugged-3da06fb1-6740-4ca1-9765-667fcac56116 for instance with vm_state building and task_state spawning. [ 868.513773] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243941, 'name': ReconfigVM_Task, 'duration_secs': 0.701892} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.514078] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 868.514280] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.514580] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-845fd1c9-25bf-4ffd-aad9-275f5eef4447 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.522518] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 868.522518] env[61907]: value = "task-1243946" [ 868.522518] env[61907]: _type = "Task" [ 868.522518] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.531329] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243946, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.731067] env[61907]: DEBUG nova.network.neutron [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.742291] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.742475] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.742823] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 868.768067] env[61907]: DEBUG nova.network.neutron [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Updated VIF entry in instance network info cache for port aa81555f-9f10-410a-b9ad-12894784fd53. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 868.769889] env[61907]: DEBUG nova.network.neutron [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Updating instance_info_cache with network_info: [{"id": "aa81555f-9f10-410a-b9ad-12894784fd53", "address": "fa:16:3e:ee:16:e6", "network": {"id": "611325fa-e3e1-47f5-be0c-1495e46ed5b2", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1884102614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc568a1103574dd498a0e783b945fc78", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "91c1da19-ab68-4127-bacd-accbaff19651", "external-id": "nsx-vlan-transportzone-319", "segmentation_id": 319, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa81555f-9f", "ovs_interfaceid": "aa81555f-9f10-410a-b9ad-12894784fd53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.857166] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c24d63-ee15-20f9-f03a-d1ca55b6477e, 'name': SearchDatastore_Task, 'duration_secs': 0.011949} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.857586] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.857871] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.857992] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.858134] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.858423] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.858671] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9c338ef9-a5aa-447c-ae69-3f380647ff66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.871392] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.871607] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.872970] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6fe3dc0-42a3-44e1-b789-0b9a33f1a031 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.878720] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243945, 'name': ReconfigVM_Task, 'duration_secs': 0.851133} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.879335] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.880057] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-875be7ca-548d-431b-8efe-8d94915beafa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.884367] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 868.884367] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ed227-4f72-2bd6-bebc-d2f5c7f58319" [ 868.884367] env[61907]: _type = "Task" [ 868.884367] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.889742] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 868.889742] env[61907]: value = "task-1243947" [ 868.889742] env[61907]: _type = "Task" [ 868.889742] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.897128] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ed227-4f72-2bd6-bebc-d2f5c7f58319, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.903119] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243947, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.935291] env[61907]: DEBUG nova.scheduler.client.report [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 869.035140] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243946, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.055345] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 869.087561] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.087833] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.088009] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.088212] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.088366] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.088516] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.092013] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.092013] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.092013] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.092013] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.092013] env[61907]: DEBUG nova.virt.hardware [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.092013] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22719644-e889-473f-8db4-18e36dfb895a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.102757] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02641548-cc10-4dab-9726-a4a387927c1f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.236659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.236659] env[61907]: DEBUG nova.objects.instance [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'flavor' on Instance uuid 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 869.270562] env[61907]: DEBUG oslo_concurrency.lockutils [req-926c12bb-2141-4bfa-9f1d-135edc356ba6 req-935e6b65-c2c3-43e1-96df-81669f79aeeb service nova] Releasing lock "refresh_cache-aec01156-431b-447e-88ef-cb8d254c6341" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.291423] env[61907]: DEBUG nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Successfully updated port: b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 869.294626] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 869.397680] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ed227-4f72-2bd6-bebc-d2f5c7f58319, 'name': SearchDatastore_Task, 'duration_secs': 0.011338} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.398821] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3ac16ad-874a-407c-9d47-552f830a0ee5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.403959] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243947, 'name': Rename_Task, 'duration_secs': 0.278166} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.404543] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.404699] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ebc581e8-c0ad-4109-8433-a43442683c86 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.407375] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 869.407375] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f3872-7283-d81a-fd3d-e11b5b6c5b26" [ 869.407375] env[61907]: _type = "Task" [ 869.407375] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.415466] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 869.415466] env[61907]: value = "task-1243948" [ 869.415466] env[61907]: _type = "Task" [ 869.415466] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.419166] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f3872-7283-d81a-fd3d-e11b5b6c5b26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.426994] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.442206] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.442636] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 869.445782] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.301s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.447674] env[61907]: INFO nova.compute.claims [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.454723] env[61907]: DEBUG nova.network.neutron [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Updating instance_info_cache with network_info: [{"id": "3da06fb1-6740-4ca1-9765-667fcac56116", "address": "fa:16:3e:55:48:48", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da06fb1-67", "ovs_interfaceid": "3da06fb1-6740-4ca1-9765-667fcac56116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.535901] env[61907]: DEBUG oslo_vmware.api [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243946, 'name': PowerOnVM_Task, 'duration_secs': 0.653633} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.536564] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.536564] env[61907]: DEBUG nova.compute.manager [None req-e7e58efc-1575-4f1a-a658-6ca5097fb5b4 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 869.537626] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb465cd9-3cbc-4373-80c8-96bd49ed61e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.746331] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87787f38-616c-4f1b-9613-1257d6431335 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.770056] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 869.770441] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dae4822e-f766-4d9c-be9e-c625fc815205 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.780302] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 869.780302] env[61907]: value = "task-1243949" [ 869.780302] env[61907]: _type = "Task" [ 869.780302] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.790076] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243949, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.793993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.794294] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.794557] env[61907]: DEBUG nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 869.923880] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521f3872-7283-d81a-fd3d-e11b5b6c5b26, 'name': SearchDatastore_Task, 'duration_secs': 0.012771} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.928281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.928611] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] aec01156-431b-447e-88ef-cb8d254c6341/aec01156-431b-447e-88ef-cb8d254c6341.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.929021] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b24847cb-2c63-40a1-be64-f4cef593c4e1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.940738] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243948, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.942529] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 869.942529] env[61907]: value = "task-1243950" [ 869.942529] env[61907]: _type = "Task" [ 869.942529] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.954360] env[61907]: DEBUG nova.compute.utils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 869.958117] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243950, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.958907] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 869.959096] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 869.961444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.961870] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Instance network_info: |[{"id": "3da06fb1-6740-4ca1-9765-667fcac56116", "address": "fa:16:3e:55:48:48", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da06fb1-67", "ovs_interfaceid": "3da06fb1-6740-4ca1-9765-667fcac56116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 869.962479] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:55:48:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '744515ee-aa5b-4c23-b959-b56c51da6b86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3da06fb1-6740-4ca1-9765-667fcac56116', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 869.972598] env[61907]: DEBUG oslo.service.loopingcall [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.974058] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 869.974779] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4166166-4c0c-467d-8ce0-95a736618dc5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.002264] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 870.002264] env[61907]: value = "task-1243951" [ 870.002264] env[61907]: _type = "Task" [ 870.002264] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.012778] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243951, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.025634] env[61907]: DEBUG nova.policy [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 870.294504] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243949, 'name': PowerOffVM_Task, 'duration_secs': 0.320924} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.296073] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.303845] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 870.310939] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4b95d41-adfd-4168-81bb-c06ed79e6cdd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.343651] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 870.343651] env[61907]: value = "task-1243952" [ 870.343651] env[61907]: _type = "Task" [ 870.343651] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.355655] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243952, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.436971] env[61907]: DEBUG oslo_vmware.api [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1243948, 'name': PowerOnVM_Task, 'duration_secs': 0.589834} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.437391] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.437847] env[61907]: INFO nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Took 9.95 seconds to spawn the instance on the hypervisor. [ 870.438243] env[61907]: DEBUG nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 870.439683] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1674faad-54e9-48fe-919c-627bc3924011 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.460328] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243950, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.513659} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.460328] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] aec01156-431b-447e-88ef-cb8d254c6341/aec01156-431b-447e-88ef-cb8d254c6341.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.460328] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.460328] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 870.465402] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ded55c7c-f498-4f18-b1c0-41714a2869bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.474722] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 870.474722] env[61907]: value = "task-1243953" [ 870.474722] env[61907]: _type = "Task" [ 870.474722] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.485071] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243953, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.513907] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243951, 'name': CreateVM_Task, 'duration_secs': 0.474506} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.514678] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 870.515435] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 870.515617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 870.515944] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 870.516562] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa605e25-0c20-48af-8c13-79c3ed2ec1c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.525688] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 870.525688] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5263f70e-9b55-4536-913a-a7aec135f241" [ 870.525688] env[61907]: _type = "Task" [ 870.525688] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.536550] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5263f70e-9b55-4536-913a-a7aec135f241, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.625088] env[61907]: WARNING nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] 9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b already exists in list: networks containing: ['9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b']. ignoring it [ 870.625088] env[61907]: WARNING nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] 9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b already exists in list: networks containing: ['9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b']. ignoring it [ 870.715741] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Successfully updated port: 9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 870.780656] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b37c2d70-a096-44e6-9611-beda80f8b4ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.795925] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69c65ed-a6aa-4c60-89ec-4d1ac3aeffde {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.829358] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Successfully created port: 6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 870.832639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be9b7b3-d469-4317-bfcb-c812394ccd68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.843945] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47513cbb-9633-473e-b2ae-d1e53c3f79c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.858326] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243952, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.866292] env[61907]: DEBUG nova.compute.provider_tree [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.961246] env[61907]: INFO nova.compute.manager [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Took 41.80 seconds to build instance. [ 870.986433] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243953, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.266396} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.986719] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.987509] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b79d1118-ad04-45b6-804c-6c1cbbe79409 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.018134] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] aec01156-431b-447e-88ef-cb8d254c6341/aec01156-431b-447e-88ef-cb8d254c6341.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 871.021291] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-640c6432-ac80-4b5e-9549-749eda82da79 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.048313] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5263f70e-9b55-4536-913a-a7aec135f241, 'name': SearchDatastore_Task, 'duration_secs': 0.009728} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.049797] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.050095] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 871.050369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.050529] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.050731] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 871.051095] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 871.051095] env[61907]: value = "task-1243954" [ 871.051095] env[61907]: _type = "Task" [ 871.051095] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.051735] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a3e81bcd-3258-4464-a38d-d1147d1aef76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.064155] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243954, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.065441] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 871.065656] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 871.066364] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-260c1ffb-0cd5-414f-a6f2-156ef0fd9e13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.072347] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 871.072347] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4179-aa02-dd72-5371-840365407a98" [ 871.072347] env[61907]: _type = "Task" [ 871.072347] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.081332] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4179-aa02-dd72-5371-840365407a98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.219069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.219215] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.219405] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 871.349198] env[61907]: DEBUG nova.network.neutron [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "address": "fa:16:3e:9f:37:bf", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c78d5e6-cc", "ovs_interfaceid": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "address": "fa:16:3e:ea:e2:90", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ff98c3-b1", "ovs_interfaceid": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.359809] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243952, 'name': ReconfigVM_Task, 'duration_secs': 0.593292} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.360123] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 871.360327] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 871.360906] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-119ba99f-8659-47eb-8cca-628d3f1a3d01 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.369748] env[61907]: DEBUG nova.scheduler.client.report [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 871.372973] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 871.372973] env[61907]: value = "task-1243955" [ 871.372973] env[61907]: _type = "Task" [ 871.372973] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.383352] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243955, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.462996] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ebdffdb2-d7da-4327-ace0-da52f6332352 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.380s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.475164] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 871.505593] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.505872] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.506053] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.506557] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.506557] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.506557] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.506762] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.507020] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.507100] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.507276] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.507457] env[61907]: DEBUG nova.virt.hardware [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.509246] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b65dc5a2-f955-41a1-88e0-2e11b5b3de3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.519903] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31a40ff1-45e8-4ee8-bc53-aea55d0e5436 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.560384] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Received event network-changed-3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 871.560941] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Refreshing instance network info cache due to event network-changed-3da06fb1-6740-4ca1-9765-667fcac56116. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 871.560941] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Acquiring lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.561054] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Acquired lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.561186] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Refreshing network info cache for port 3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 871.568727] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243954, 'name': ReconfigVM_Task, 'duration_secs': 0.285803} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.569522] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Reconfigured VM instance instance-00000049 to attach disk [datastore2] aec01156-431b-447e-88ef-cb8d254c6341/aec01156-431b-447e-88ef-cb8d254c6341.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.570155] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-858cc5eb-01e6-4fa1-ba54-f83f1cbcb883 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.578447] env[61907]: DEBUG nova.compute.manager [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Received event network-vif-plugged-9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 871.578691] env[61907]: DEBUG oslo_concurrency.lockutils [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] Acquiring lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.578922] env[61907]: DEBUG oslo_concurrency.lockutils [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] Lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.579115] env[61907]: DEBUG oslo_concurrency.lockutils [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] Lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.579343] env[61907]: DEBUG nova.compute.manager [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] No waiting events found dispatching network-vif-plugged-9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 871.579623] env[61907]: WARNING nova.compute.manager [req-f494a4ea-f989-4577-bcb6-b90cdb47329f req-86416a92-f3be-4825-a683-a4e0cdb3d835 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Received unexpected event network-vif-plugged-9897ff2a-3de3-42a2-af12-5f0ef6234904 for instance with vm_state building and task_state spawning. [ 871.583644] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 871.583644] env[61907]: value = "task-1243956" [ 871.583644] env[61907]: _type = "Task" [ 871.583644] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.587360] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be4179-aa02-dd72-5371-840365407a98, 'name': SearchDatastore_Task, 'duration_secs': 0.01165} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.591826] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7bd80ef-b089-4b02-a79f-476ed1a29198 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.597676] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 871.597676] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52518760-f5bd-427e-f0b8-377d710de732" [ 871.597676] env[61907]: _type = "Task" [ 871.597676] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.602143] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243956, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.615770] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52518760-f5bd-427e-f0b8-377d710de732, 'name': SearchDatastore_Task, 'duration_secs': 0.011996} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.616146] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.616454] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] d32315e1-68b9-49ee-9393-8b2f9cbaf645/d32315e1-68b9-49ee-9393-8b2f9cbaf645.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 871.616793] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b5ebce7-1369-47ea-813d-4c3978239ffd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.625800] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 871.625800] env[61907]: value = "task-1243957" [ 871.625800] env[61907]: _type = "Task" [ 871.625800] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.634748] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.771033] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 871.855700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.856450] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.856628] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.857626] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfab434b-0789-4906-a5a7-896169d2cc48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.882150] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.435s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.882546] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 871.882677] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 871.882747] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 871.883526] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 871.883526] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 871.883526] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 871.883769] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 871.883769] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 871.885536] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 871.885536] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 871.885536] env[61907]: DEBUG nova.virt.hardware [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 871.891548] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfiguring VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 871.895480] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.367s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.895691] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.895853] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 871.896210] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.860s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.896408] env[61907]: DEBUG nova.objects.instance [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lazy-loading 'pci_requests' on Instance uuid 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.898380] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-84f30538-544b-4f41-a19d-f124d26d0a27 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.915030] env[61907]: DEBUG nova.objects.instance [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lazy-loading 'numa_topology' on Instance uuid 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 871.916598] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66307e95-f1ca-48a9-8040-408289240dd1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.934431] env[61907]: DEBUG oslo_vmware.api [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1243955, 'name': PowerOnVM_Task, 'duration_secs': 0.5224} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.936870] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f6cbb00-05fb-4288-b77a-8ba0346c8f2a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.942478] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 871.942478] env[61907]: DEBUG nova.compute.manager [None req-bf050e08-de69-417d-9aca-b7751caf49e1 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 871.942777] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 871.942777] env[61907]: value = "task-1243958" [ 871.942777] env[61907]: _type = "Task" [ 871.942777] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.943921] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9f0295-a5e0-4960-b647-9f881355c0e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.963666] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943ac53f-a67f-400a-ac8d-839369607eb0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.975361] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243958, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.981681] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26af88fa-d02c-496a-a587-8ce1c407d928 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.014953] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180385MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 872.015171] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.043723] env[61907]: DEBUG nova.network.neutron [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Updating instance_info_cache with network_info: [{"id": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "address": "fa:16:3e:82:95:d5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897ff2a-3d", "ovs_interfaceid": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.104058] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243956, 'name': Rename_Task, 'duration_secs': 0.203405} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.104938] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.105409] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-722de4aa-af8a-4d39-9a27-67036b973963 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.117018] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 872.117018] env[61907]: value = "task-1243959" [ 872.117018] env[61907]: _type = "Task" [ 872.117018] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.130543] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243959, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.144788] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243957, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.393537] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "e91ef672-48ef-4d54-a01b-428b703132a0" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.393839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "e91ef672-48ef-4d54-a01b-428b703132a0" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.422686] env[61907]: INFO nova.compute.claims [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.441540] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Updated VIF entry in instance network info cache for port 3da06fb1-6740-4ca1-9765-667fcac56116. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 872.441915] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Updating instance_info_cache with network_info: [{"id": "3da06fb1-6740-4ca1-9765-667fcac56116", "address": "fa:16:3e:55:48:48", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3da06fb1-67", "ovs_interfaceid": "3da06fb1-6740-4ca1-9765-667fcac56116", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.459207] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243958, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.547814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.547814] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Instance network_info: |[{"id": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "address": "fa:16:3e:82:95:d5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897ff2a-3d", "ovs_interfaceid": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 872.548124] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:95:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9897ff2a-3de3-42a2-af12-5f0ef6234904', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 872.557717] env[61907]: DEBUG oslo.service.loopingcall [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.557962] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 872.558322] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fedb3888-c902-42a9-97d2-1a82bd6c96ec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.584829] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 872.584829] env[61907]: value = "task-1243960" [ 872.584829] env[61907]: _type = "Task" [ 872.584829] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.596377] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243960, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.628164] env[61907]: DEBUG oslo_vmware.api [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243959, 'name': PowerOnVM_Task, 'duration_secs': 0.487065} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.628164] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 872.628164] env[61907]: INFO nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Took 8.65 seconds to spawn the instance on the hypervisor. [ 872.628164] env[61907]: DEBUG nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 872.629182] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5dbfc3-2f77-4eaa-bbf6-bc9602978eea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.645731] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.558537} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.647308] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] d32315e1-68b9-49ee-9393-8b2f9cbaf645/d32315e1-68b9-49ee-9393-8b2f9cbaf645.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 872.647572] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 872.652730] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d02cd33d-0d7a-4463-808c-b38330e99ab8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.654534] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Successfully updated port: 6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 872.662499] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 872.662499] env[61907]: value = "task-1243961" [ 872.662499] env[61907]: _type = "Task" [ 872.662499] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.674443] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243961, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.783166] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.783563] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.783879] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.784251] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.784599] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.789213] env[61907]: INFO nova.compute.manager [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Terminating instance [ 872.896154] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "e91ef672-48ef-4d54-a01b-428b703132a0" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.896760] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 872.944874] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Releasing lock "refresh_cache-d32315e1-68b9-49ee-9393-8b2f9cbaf645" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.945202] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-plugged-b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 872.945405] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.945609] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.945770] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.945954] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] No waiting events found dispatching network-vif-plugged-b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 872.946149] env[61907]: WARNING nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received unexpected event network-vif-plugged-b0ff98c3-b17f-497e-8ba2-0578ebbd2927 for instance with vm_state active and task_state None. [ 872.946316] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-changed-b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 872.946471] env[61907]: DEBUG nova.compute.manager [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing instance network info cache due to event network-changed-b0ff98c3-b17f-497e-8ba2-0578ebbd2927. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 872.946653] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.946800] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.947035] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Refreshing network info cache for port b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 872.959936] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243958, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.101544] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243960, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.161582] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.161582] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.161753] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 873.171582] env[61907]: INFO nova.compute.manager [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Took 38.94 seconds to build instance. [ 873.179685] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243961, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.166959} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.180503] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 873.181634] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e74b386-4e68-4051-89f0-267b3a5dd5be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.212252] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] d32315e1-68b9-49ee-9393-8b2f9cbaf645/d32315e1-68b9-49ee-9393-8b2f9cbaf645.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 873.215535] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4d55f1b-cc78-4db6-9e8f-f8c361e59137 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.240069] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 873.240069] env[61907]: value = "task-1243962" [ 873.240069] env[61907]: _type = "Task" [ 873.240069] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.252614] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243962, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.277919] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a05681-b492-4bd7-88c4-b085aca099fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.286584] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066e8e62-67eb-49b2-b79e-b5b9871ab2ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.292625] env[61907]: DEBUG nova.compute.manager [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 873.292900] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.293915] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1088d75f-99c2-41b1-a697-11e0c242b983 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.324997] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58659b3-8fb5-4f6e-ab5c-289795f1273d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.330651] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.331479] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc09cb45-2c63-4822-8ef3-295a837526f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.337307] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa9ff66-6191-404f-a37f-84fd4e83421d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.343298] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 873.343298] env[61907]: value = "task-1243963" [ 873.343298] env[61907]: _type = "Task" [ 873.343298] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.360918] env[61907]: DEBUG nova.compute.provider_tree [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.369908] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243963, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.403425] env[61907]: DEBUG nova.compute.utils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.405067] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 873.405268] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.462868] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243958, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.498774] env[61907]: DEBUG nova.policy [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae271fea344342b0b2027643f01c929a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4cce4c7593147bb8fc26486807edc20', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.599727] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243960, 'name': CreateVM_Task, 'duration_secs': 0.725146} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.599950] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.600944] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.601399] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.601835] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.602342] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7b0713-ecc8-446f-81fc-f69a2ecbee5f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.609727] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 873.609727] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c64ff4-ee9c-1212-266d-d4af3793d46d" [ 873.609727] env[61907]: _type = "Task" [ 873.609727] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.619868] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c64ff4-ee9c-1212-266d-d4af3793d46d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.674212] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a23db325-4671-4e7c-b042-a29a60ff2193 tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.275s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.754340] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243962, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.755265] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 873.854745] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243963, 'name': PowerOffVM_Task, 'duration_secs': 0.322424} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.855040] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.855388] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.855688] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb439f4a-56aa-49bf-a705-6ed7e42c7de3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.864377] env[61907]: DEBUG nova.scheduler.client.report [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 873.909229] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 873.939893] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.940137] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.940418] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleting the datastore file [datastore2] ddcbf498-fd27-41f2-bbcc-23a8f9827823 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.940730] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2647dc23-2212-4655-81c0-d480151f29c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.950697] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 873.950697] env[61907]: value = "task-1243965" [ 873.950697] env[61907]: _type = "Task" [ 873.950697] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.967972] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243965, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.971248] env[61907]: DEBUG oslo_vmware.api [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243958, 'name': ReconfigVM_Task, 'duration_secs': 1.802756} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.972019] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.972250] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfigured VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 873.993661] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updated VIF entry in instance network info cache for port b0ff98c3-b17f-497e-8ba2-0578ebbd2927. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 873.994051] env[61907]: DEBUG nova.network.neutron [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "address": "fa:16:3e:9f:37:bf", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c78d5e6-cc", "ovs_interfaceid": "0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "address": "fa:16:3e:ea:e2:90", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ff98c3-b1", "ovs_interfaceid": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.054156] env[61907]: DEBUG nova.network.neutron [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Updating instance_info_cache with network_info: [{"id": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "address": "fa:16:3e:2a:ef:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ef9178a-ce", "ovs_interfaceid": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.124021] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c64ff4-ee9c-1212-266d-d4af3793d46d, 'name': SearchDatastore_Task, 'duration_secs': 0.01925} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.124021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.124021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.124021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.124021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.124021] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.124021] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fce0b45-6a6e-4908-8024-ff4d1fe1bd59 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.134465] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.134738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.135702] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84161448-83af-4a33-bd52-ba1096c44ecf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.140975] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Successfully created port: d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.148050] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 874.148050] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524205d3-ed43-1e0b-72ca-860beffb1ccf" [ 874.148050] env[61907]: _type = "Task" [ 874.148050] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.157321] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524205d3-ed43-1e0b-72ca-860beffb1ccf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.254248] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243962, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.362450] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Received event network-changed-9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 874.362725] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Refreshing instance network info cache due to event network-changed-9897ff2a-3de3-42a2-af12-5f0ef6234904. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 874.363480] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Acquiring lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.363767] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Acquired lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.364009] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Refreshing network info cache for port 9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 874.370715] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.373142] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.358s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.460332] env[61907]: INFO nova.network.neutron [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating port 40a79714-91a5-4175-94fe-d7aeae0a8ea5 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 874.468572] env[61907]: DEBUG oslo_vmware.api [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243965, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277733} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.468863] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.468934] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.469125] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.469304] env[61907]: INFO nova.compute.manager [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Took 1.18 seconds to destroy the instance on the hypervisor. [ 874.469546] env[61907]: DEBUG oslo.service.loopingcall [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.469741] env[61907]: DEBUG nova.compute.manager [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 874.469838] env[61907]: DEBUG nova.network.neutron [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.478130] env[61907]: DEBUG oslo_concurrency.lockutils [None req-efe3c002-d3e9-4001-b0ca-08cd07fae9e2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-b0ff98c3-b17f-497e-8ba2-0578ebbd2927" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 9.294s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.496581] env[61907]: DEBUG oslo_concurrency.lockutils [req-e049b6f2-996f-47a9-84d2-2fdebe4cdd00 req-a2ac096f-de24-45d6-a3bd-7c73e659449b service nova] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.558375] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.558708] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Instance network_info: |[{"id": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "address": "fa:16:3e:2a:ef:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ef9178a-ce", "ovs_interfaceid": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 874.559178] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:ef:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ef9178a-ce1e-4b38-ba1d-398b51e22951', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.572372] env[61907]: DEBUG oslo.service.loopingcall [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.572372] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.572372] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a4dc03c9-6466-407e-b9a0-ca4ce0db25f9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.605327] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.605327] env[61907]: value = "task-1243966" [ 874.605327] env[61907]: _type = "Task" [ 874.605327] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.620310] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243966, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.660525] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524205d3-ed43-1e0b-72ca-860beffb1ccf, 'name': SearchDatastore_Task, 'duration_secs': 0.013143} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.661376] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ee20733-cb70-4edc-92ca-2f81e3766f8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.667859] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 874.667859] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fd9f52-01aa-6bad-f9d1-3d704435168c" [ 874.667859] env[61907]: _type = "Task" [ 874.667859] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.678675] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fd9f52-01aa-6bad-f9d1-3d704435168c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.753148] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243962, 'name': ReconfigVM_Task, 'duration_secs': 1.508966} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.753497] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Reconfigured VM instance instance-0000004a to attach disk [datastore2] d32315e1-68b9-49ee-9393-8b2f9cbaf645/d32315e1-68b9-49ee-9393-8b2f9cbaf645.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 874.754213] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c82403ad-99fe-4547-af7e-217535d1c2b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.763265] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 874.763265] env[61907]: value = "task-1243967" [ 874.763265] env[61907]: _type = "Task" [ 874.763265] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.774145] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243967, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.920156] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 874.950573] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.950877] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.951179] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.951469] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.951697] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.951901] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.952186] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.952423] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.952643] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.952937] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.953170] env[61907]: DEBUG nova.virt.hardware [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.954180] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd100335-5988-45ee-8c2f-563904346553 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.963687] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07255893-74d2-416c-841a-555e6c971ad9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.117673] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243966, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.179772] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fd9f52-01aa-6bad-f9d1-3d704435168c, 'name': SearchDatastore_Task, 'duration_secs': 0.026572} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.180144] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.180458] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e52648ac-e068-4a63-baa0-a1c34df52197/e52648ac-e068-4a63-baa0-a1c34df52197.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.180771] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-90294f7e-4411-4e42-bf82-cf046cb36502 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.188770] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 875.188770] env[61907]: value = "task-1243968" [ 875.188770] env[61907]: _type = "Task" [ 875.188770] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.197695] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243968, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.198639] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Updated VIF entry in instance network info cache for port 9897ff2a-3de3-42a2-af12-5f0ef6234904. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 875.199046] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Updating instance_info_cache with network_info: [{"id": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "address": "fa:16:3e:82:95:d5", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9897ff2a-3d", "ovs_interfaceid": "9897ff2a-3de3-42a2-af12-5f0ef6234904", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.275066] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243967, 'name': Rename_Task, 'duration_secs': 0.206633} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.275374] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 875.275634] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c87ace4e-7141-4738-88cc-386ad7e206bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.283848] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 875.283848] env[61907]: value = "task-1243969" [ 875.283848] env[61907]: _type = "Task" [ 875.283848] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.295774] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.417015] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 865d386d-c5f0-439b-b50b-b613993ec9b8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417015] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 52404147-0375-4f9c-9e1e-4a52001df7cb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417015] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417228] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417261] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23526967-cd8f-4581-b9c5-1c270d385163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417371] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance da490943-511b-4776-8f16-4f51c3b055c4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417498] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417614] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ddcbf498-fd27-41f2-bbcc-23a8f9827823 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417726] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 41c680fb-5450-43f3-9acb-5218e57a7da9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.417836] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f028beaf-4185-40d7-a730-abe91209f9d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2c1212a7-df13-49b7-b3b3-24fb59b6789d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance aec01156-431b-447e-88ef-cb8d254c6341 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance d32315e1-68b9-49ee-9393-8b2f9cbaf645 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e52648ac-e068-4a63-baa0-a1c34df52197 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 37ff6c54-6b79-4a9c-bc16-29d974185d3e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.418728] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5c357582-0b55-4de7-968e-4daa3522dda4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 875.419497] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 18 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 875.419855] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=4032MB phys_disk=200GB used_disk=18GB total_vcpus=48 used_vcpus=18 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 875.626023] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243966, 'name': CreateVM_Task, 'duration_secs': 0.54741} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.626463] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.627677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.627908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.628211] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.628487] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a13a5d75-46a5-4290-83e0-32462b2468c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.637774] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 875.637774] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52021c0d-0f54-7860-dd9b-5ae91d88bdf4" [ 875.637774] env[61907]: _type = "Task" [ 875.637774] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.647452] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52021c0d-0f54-7860-dd9b-5ae91d88bdf4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.703484] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Releasing lock "refresh_cache-e52648ac-e068-4a63-baa0-a1c34df52197" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.703771] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Received event network-vif-plugged-6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 875.704223] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Acquiring lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.704352] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.704725] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.704725] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] No waiting events found dispatching network-vif-plugged-6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 875.704850] env[61907]: WARNING nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Received unexpected event network-vif-plugged-6ef9178a-ce1e-4b38-ba1d-398b51e22951 for instance with vm_state building and task_state spawning. [ 875.705121] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Received event network-changed-6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 875.705369] env[61907]: DEBUG nova.compute.manager [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Refreshing instance network info cache due to event network-changed-6ef9178a-ce1e-4b38-ba1d-398b51e22951. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 875.705578] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Acquiring lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.705728] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Acquired lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.706098] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Refreshing network info cache for port 6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.709366] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243968, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.793020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942a83d0-5520-4df5-9fd9-364a241e544b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.807919] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fe4c380-093f-4808-a400-be3b5c744984 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.811381] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.846639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f0c88c-e999-4136-9e43-d4cb90549615 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.855045] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96af5aa0-85e6-41e7-8747-a2c5713068ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.868736] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.925719] env[61907]: DEBUG nova.network.neutron [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.149144] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52021c0d-0f54-7860-dd9b-5ae91d88bdf4, 'name': SearchDatastore_Task, 'duration_secs': 0.057016} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.149510] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.149793] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.150084] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.150340] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.150578] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.150889] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a4acce5b-4045-43b5-a061-2b66c00c13dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.164803] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.165290] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 876.165876] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cba9bb83-502e-42e2-b561-169e46c3d814 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.171486] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 876.171486] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f930d9-38e6-07fd-6495-d070c7c10037" [ 876.171486] env[61907]: _type = "Task" [ 876.171486] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.179851] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f930d9-38e6-07fd-6495-d070c7c10037, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.199500] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243968, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.302139] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243969, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.371488] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 876.428250] env[61907]: INFO nova.compute.manager [-] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Took 1.96 seconds to deallocate network for instance. [ 876.613262] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Successfully updated port: d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.687925] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f930d9-38e6-07fd-6495-d070c7c10037, 'name': SearchDatastore_Task, 'duration_secs': 0.03046} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.691287] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1af8ba68-e169-4979-9f6f-08e210b7da16 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.703300] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 876.703300] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c19e3-06c7-3d48-2fa1-d28e54b2fbbe" [ 876.703300] env[61907]: _type = "Task" [ 876.703300] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.704945] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243968, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.713226] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c19e3-06c7-3d48-2fa1-d28e54b2fbbe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.805481] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243969, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.818634] env[61907]: DEBUG nova.compute.manager [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Received event network-vif-deleted-9f84bd0f-4a0a-43d1-b913-e6667f455b55 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 876.818961] env[61907]: DEBUG nova.compute.manager [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Received event network-vif-plugged-d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 876.819987] env[61907]: DEBUG oslo_concurrency.lockutils [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] Acquiring lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.820243] env[61907]: DEBUG oslo_concurrency.lockutils [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] Lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.820570] env[61907]: DEBUG oslo_concurrency.lockutils [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] Lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.820817] env[61907]: DEBUG nova.compute.manager [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] No waiting events found dispatching network-vif-plugged-d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 876.821108] env[61907]: WARNING nova.compute.manager [req-f67d9df5-b5a3-4198-8701-0c6ea142872e req-de23d52b-1bce-433f-a68f-f21e2be16f58 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Received unexpected event network-vif-plugged-d1082c95-320a-4bba-b83c-c3e5da0d2384 for instance with vm_state building and task_state spawning. [ 876.841042] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Updated VIF entry in instance network info cache for port 6ef9178a-ce1e-4b38-ba1d-398b51e22951. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.841483] env[61907]: DEBUG nova.network.neutron [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Updating instance_info_cache with network_info: [{"id": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "address": "fa:16:3e:2a:ef:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ef9178a-ce", "ovs_interfaceid": "6ef9178a-ce1e-4b38-ba1d-398b51e22951", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.877330] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 876.877330] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.504s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.938333] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.938718] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.938963] env[61907]: DEBUG nova.objects.instance [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lazy-loading 'resources' on Instance uuid ddcbf498-fd27-41f2-bbcc-23a8f9827823 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.117295] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.117570] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquired lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.118919] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.204731] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243968, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.632822} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.205124] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] e52648ac-e068-4a63-baa0-a1c34df52197/e52648ac-e068-4a63-baa0-a1c34df52197.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.205366] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.210255] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed819365-53b6-4f1d-bba8-27f9b657e93c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.218474] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c19e3-06c7-3d48-2fa1-d28e54b2fbbe, 'name': SearchDatastore_Task, 'duration_secs': 0.043499} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.219781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.220144] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 37ff6c54-6b79-4a9c-bc16-29d974185d3e/37ff6c54-6b79-4a9c-bc16-29d974185d3e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 877.221139] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 877.221139] env[61907]: value = "task-1243970" [ 877.221139] env[61907]: _type = "Task" [ 877.221139] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.221448] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03d5edf0-659e-49dc-bf1d-b39b88d871f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.234912] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.236404] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 877.236404] env[61907]: value = "task-1243971" [ 877.236404] env[61907]: _type = "Task" [ 877.236404] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.245110] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.297549] env[61907]: DEBUG oslo_vmware.api [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243969, 'name': PowerOnVM_Task, 'duration_secs': 1.840376} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.297910] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 877.298096] env[61907]: INFO nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Took 10.68 seconds to spawn the instance on the hypervisor. [ 877.298339] env[61907]: DEBUG nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.300594] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11a95302-bfa1-40a8-8c73-c651cb13f795 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.346917] env[61907]: DEBUG oslo_concurrency.lockutils [req-a60dfd5c-9d1a-4676-917c-6d977a0a598d req-983d8af9-c912-4709-b073-eaf03bd616a1 service nova] Releasing lock "refresh_cache-37ff6c54-6b79-4a9c-bc16-29d974185d3e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.360908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.360908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.360908] env[61907]: DEBUG nova.network.neutron [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.366588] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 877.366835] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 877.423341] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.423616] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.423807] env[61907]: DEBUG nova.compute.manager [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 877.424973] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc678ac1-0b63-40cf-b827-1450ac8003db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.433802] env[61907]: DEBUG nova.compute.manager [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 877.434580] env[61907]: DEBUG nova.objects.instance [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'flavor' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.690508] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.756705] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068155} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.762641] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.763369] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469254} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.765998] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a3bf9c-c67f-499e-8081-ea501fa41d51 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.768981] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 37ff6c54-6b79-4a9c-bc16-29d974185d3e/37ff6c54-6b79-4a9c-bc16-29d974185d3e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.769229] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.769804] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f144688-1a89-4bc6-8f9c-59b081bcfdcc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.796022] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] e52648ac-e068-4a63-baa0-a1c34df52197/e52648ac-e068-4a63-baa0-a1c34df52197.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.798742] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e0ed4669-35bc-4a0b-9622-a08f62fcf68b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.815597] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 877.815597] env[61907]: value = "task-1243972" [ 877.815597] env[61907]: _type = "Task" [ 877.815597] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.825605] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 877.825605] env[61907]: value = "task-1243973" [ 877.825605] env[61907]: _type = "Task" [ 877.825605] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.835670] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243972, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.839017] env[61907]: INFO nova.compute.manager [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Took 39.50 seconds to build instance. [ 877.839017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "aec01156-431b-447e-88ef-cb8d254c6341" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.839017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.839017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "aec01156-431b-447e-88ef-cb8d254c6341-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.839017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.839017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.841572] env[61907]: INFO nova.compute.manager [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Terminating instance [ 877.846790] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243973, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.852397] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1372cb39-ae49-4503-8f4e-09bc3ea2ae18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.861457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489a4222-d8ec-4ac8-8bc2-c21e2a25962e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.901741] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a2daea-9a96-416f-9974-cfde7b2fc52c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.911618] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ea2f4e-6dc4-485f-91df-5bd33e43fc12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.924149] env[61907]: DEBUG nova.compute.provider_tree [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.211791] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.212245] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.212383] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 878.227197] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.227524] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.246587] env[61907]: DEBUG nova.compute.manager [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 878.331845] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243972, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.261577} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.336174] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.337370] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2a77a86-2539-45b8-98c9-5109a9f807cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.342479] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8b98cdd3-8020-450f-b415-f07ba2285cca tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.117s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.348298] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243973, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.364271] env[61907]: DEBUG nova.compute.manager [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 878.364620] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 878.379903] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 37ff6c54-6b79-4a9c-bc16-29d974185d3e/37ff6c54-6b79-4a9c-bc16-29d974185d3e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.381189] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ec1b68-c42b-436e-abd8-43249bf912fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.386428] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d4d310a-a06a-4cd9-8139-0dd9e39f4ae8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.411345] env[61907]: DEBUG nova.network.neutron [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Updating instance_info_cache with network_info: [{"id": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "address": "fa:16:3e:65:cb:86", "network": {"id": "8ad383b2-b578-47d1-8cc1-7805aae4ee55", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-293583221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4cce4c7593147bb8fc26486807edc20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1082c95-32", "ovs_interfaceid": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.422782] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.424658] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47b03374-b8ef-4b81-83e2-07977468ed92 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.429749] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 878.429749] env[61907]: value = "task-1243974" [ 878.429749] env[61907]: _type = "Task" [ 878.429749] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.429749] env[61907]: DEBUG nova.scheduler.client.report [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 878.439485] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 878.439485] env[61907]: value = "task-1243975" [ 878.439485] env[61907]: _type = "Task" [ 878.439485] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.446578] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 878.447749] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243974, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.448010] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6565c860-31d6-4ca6-a7f6-0beb63c7708f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.455261] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.456808] env[61907]: DEBUG oslo_vmware.api [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 878.456808] env[61907]: value = "task-1243976" [ 878.456808] env[61907]: _type = "Task" [ 878.456808] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.468090] env[61907]: DEBUG oslo_vmware.api [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.682308] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "41c680fb-5450-43f3-9acb-5218e57a7da9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.682691] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.682964] env[61907]: DEBUG nova.compute.manager [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 878.684020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3818b40d-0d00-4a49-940e-9caee0643719 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.697026] env[61907]: DEBUG nova.compute.manager [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 878.697026] env[61907]: DEBUG nova.objects.instance [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'flavor' on Instance uuid 41c680fb-5450-43f3-9acb-5218e57a7da9 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 878.732035] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 878.740099] env[61907]: DEBUG nova.network.neutron [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.746449] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.746449] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.773407] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.844136] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243973, 'name': ReconfigVM_Task, 'duration_secs': 0.823735} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.844451] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Reconfigured VM instance instance-0000004b to attach disk [datastore2] e52648ac-e068-4a63-baa0-a1c34df52197/e52648ac-e068-4a63-baa0-a1c34df52197.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.845169] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fdb8fe96-4e0f-442a-bf66-59287b7b86f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.851337] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 878.851337] env[61907]: value = "task-1243977" [ 878.851337] env[61907]: _type = "Task" [ 878.851337] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.859628] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243977, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.917191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Releasing lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.917191] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Instance network_info: |[{"id": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "address": "fa:16:3e:65:cb:86", "network": {"id": "8ad383b2-b578-47d1-8cc1-7805aae4ee55", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-293583221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4cce4c7593147bb8fc26486807edc20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1082c95-32", "ovs_interfaceid": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 878.917191] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:cb:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2e8b74b-aa27-4f31-9414-7bcf531e8642', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd1082c95-320a-4bba-b83c-c3e5da0d2384', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.923319] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Creating folder: Project (b4cce4c7593147bb8fc26486807edc20). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.923780] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4516525f-e9fc-4239-b7d9-3b19c1b08059 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.935203] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Created folder: Project (b4cce4c7593147bb8fc26486807edc20) in parent group-v268168. [ 878.935203] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Creating folder: Instances. Parent ref: group-v268267. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.938095] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1a0e2ef-da50-4da1-9692-a9a2e0d51735 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.940794] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.943088] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.170s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.957514] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243974, 'name': ReconfigVM_Task, 'duration_secs': 0.399729} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.962272] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 37ff6c54-6b79-4a9c-bc16-29d974185d3e/37ff6c54-6b79-4a9c-bc16-29d974185d3e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 878.963957] env[61907]: INFO nova.scheduler.client.report [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleted allocations for instance ddcbf498-fd27-41f2-bbcc-23a8f9827823 [ 878.970749] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2b93ff6-8371-48f6-b0fd-c9ca34f7f3ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.973097] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Created folder: Instances in parent group-v268267. [ 878.973447] env[61907]: DEBUG oslo.service.loopingcall [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.973780] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243975, 'name': PowerOffVM_Task, 'duration_secs': 0.22222} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.977591] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.977969] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 878.978298] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 878.979200] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b5bdf4aa-67fe-4056-8ab9-415e3bb3acd6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.998931] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-41c1e0f9-c605-4575-9679-65efd0c1eabe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.010859] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Received event network-changed-d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 879.010986] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Refreshing instance network info cache due to event network-changed-d1082c95-320a-4bba-b83c-c3e5da0d2384. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 879.011260] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Acquiring lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.011616] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Acquired lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.011869] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Refreshing network info cache for port d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.013523] env[61907]: DEBUG oslo_vmware.api [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1243976, 'name': PowerOffVM_Task, 'duration_secs': 0.385568} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.014878] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 879.014878] env[61907]: value = "task-1243980" [ 879.014878] env[61907]: _type = "Task" [ 879.014878] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.016170] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 879.016562] env[61907]: DEBUG nova.compute.manager [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 879.018123] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b11fb6e-98af-46f6-8edf-d724cc5e617c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.029226] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 879.029226] env[61907]: value = "task-1243981" [ 879.029226] env[61907]: _type = "Task" [ 879.029226] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.040647] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243980, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.046326] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243981, 'name': CreateVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.103346] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 879.103712] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 879.103980] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Deleting the datastore file [datastore2] aec01156-431b-447e-88ef-cb8d254c6341 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 879.104688] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a2131eae-7a55-4659-ae65-fd8bcf1da9a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.112581] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for the task: (returnval){ [ 879.112581] env[61907]: value = "task-1243983" [ 879.112581] env[61907]: _type = "Task" [ 879.112581] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.121960] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243983, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.243022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.248152] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.248391] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.249868] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aec4cfcd-b88b-4cb6-98cf-75d382b409f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.257240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.276314] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d14ca76-245c-470e-8eb3-7db6dc26bf4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.312484] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfiguring VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 879.315621] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0c23326a687222e63a7087617c8852e9',container_format='bare',created_at=2024-10-10T13:11:45Z,direct_url=,disk_format='vmdk',id=c525c73f-cabe-4077-be2c-3ee109d7c7c8,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-2030004455-shelved',owner='3aab40034a4e4ccdb169abfc11f3b02f',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-10T13:11:59Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.316424] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.316424] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.316599] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.316839] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.317132] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.317471] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.318058] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.318058] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.318171] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.318517] env[61907]: DEBUG nova.virt.hardware [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.319098] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39a66619-2dc7-4abe-af0c-221bc4d1f0ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.340917] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d4ae64-224f-4dca-9b77-160b9c4e25da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.140842] env[61907]: INFO nova.compute.claims [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.151037] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 880.153360] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5b6072b-7a54-4acd-b3a2-e1737c3f4304 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.730s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.167853] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-590dabad-c677-48d2-aeba-5ac015fcfb71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.176293] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 880.176293] env[61907]: value = "task-1243984" [ 880.176293] env[61907]: _type = "Task" [ 880.176293] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.176926] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1cc30943-42cc-4785-a963-bb1d132fef28 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "ddcbf498-fd27-41f2-bbcc-23a8f9827823" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.393s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.194519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c93c8e-991b-4e98-a267-c0a0f071ed8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.199136] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243977, 'name': Rename_Task, 'duration_secs': 0.329948} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.207891] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.208542] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243980, 'name': Rename_Task, 'duration_secs': 0.149334} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.208776] env[61907]: DEBUG oslo_vmware.api [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Task: {'id': task-1243983, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184038} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.212859] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-672e6837-2e00-483c-bd6c-43f1ae156d2d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.214634] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.215613] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.215613] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 880.215613] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 880.215613] env[61907]: INFO nova.compute.manager [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Took 1.85 seconds to destroy the instance on the hypervisor. [ 880.215749] env[61907]: DEBUG oslo.service.loopingcall [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.216170] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243981, 'name': CreateVM_Task, 'duration_secs': 0.596326} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.232284] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1423725d-c8be-4c4b-8521-8e4c4ff4ccad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.232284] env[61907]: DEBUG nova.compute.manager [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 880.232284] env[61907]: DEBUG nova.network.neutron [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 880.233034] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.233404] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.234789] env[61907]: DEBUG oslo_vmware.api [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 880.234789] env[61907]: value = "task-1243985" [ 880.234789] env[61907]: _type = "Task" [ 880.234789] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.234975] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:94:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40a79714-91a5-4175-94fe-d7aeae0a8ea5', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.242170] env[61907]: DEBUG oslo.service.loopingcall [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.243348] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.243514] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.243842] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.246651] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.247339] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cef42f37-8a28-429e-b03c-0690d5ccfa57 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.252226] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aeccb0b6-5bd3-4a12-a6fc-a4af8d148f46 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.271312] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 880.271312] env[61907]: value = "task-1243986" [ 880.271312] env[61907]: _type = "Task" [ 880.271312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.271312] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 880.271312] env[61907]: value = "task-1243987" [ 880.271312] env[61907]: _type = "Task" [ 880.271312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.287423] env[61907]: DEBUG oslo_vmware.api [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243985, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.287723] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.287723] env[61907]: value = "task-1243988" [ 880.287723] env[61907]: _type = "Task" [ 880.287723] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.288352] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 880.288352] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525a344e-528c-54fb-b4a3-1a47abe47d64" [ 880.288352] env[61907]: _type = "Task" [ 880.288352] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.302469] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243987, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.303278] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243986, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.309143] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243988, 'name': CreateVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.312742] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525a344e-528c-54fb-b4a3-1a47abe47d64, 'name': SearchDatastore_Task, 'duration_secs': 0.017799} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.313135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.313531] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.313969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.313969] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.314134] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.314488] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0609683e-82db-48b5-ae81-fa8bf9f143ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.325286] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.325624] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.326637] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef4a714f-a1a3-4876-a519-5dcaf0d933f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.333131] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 880.333131] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52db7a1b-379b-5d0a-98af-f86b60aa8756" [ 880.333131] env[61907]: _type = "Task" [ 880.333131] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.343602] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52db7a1b-379b-5d0a-98af-f86b60aa8756, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.658208] env[61907]: INFO nova.compute.resource_tracker [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating resource usage from migration 50cbf55f-4d6f-40b0-af2d-1eaddbc36258 [ 880.674703] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "865d386d-c5f0-439b-b50b-b613993ec9b8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.675023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.675302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.675540] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.675753] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.681681] env[61907]: INFO nova.compute.manager [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Terminating instance [ 880.706209] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.751124] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [{"id": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "address": "fa:16:3e:0f:68:53", "network": {"id": "db75e4c7-0a50-4d20-9c0c-fb11ae241f1b", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1226576763-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "8e6a77496c4f4c7ea060d39ad1504c53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bfa7abe-7e46-4d8f-b50a-4d0c4509e4dc", "external-id": "nsx-vlan-transportzone-951", "segmentation_id": 951, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap05a00a60-c2", "ovs_interfaceid": "05a00a60-c277-47b8-98e8-bc8f0f46258a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.763226] env[61907]: DEBUG oslo_vmware.api [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1243985, 'name': PowerOffVM_Task, 'duration_secs': 0.274437} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.763498] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 880.763723] env[61907]: DEBUG nova.compute.manager [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 880.764613] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f2272f-df0b-4e70-9bb1-1e64ecbeed1c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.805802] env[61907]: DEBUG oslo_vmware.api [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243986, 'name': PowerOnVM_Task, 'duration_secs': 0.536283} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.809267] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.809503] env[61907]: INFO nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Took 11.75 seconds to spawn the instance on the hypervisor. [ 880.809723] env[61907]: DEBUG nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 880.810033] env[61907]: DEBUG oslo_vmware.api [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243987, 'name': PowerOnVM_Task, 'duration_secs': 0.539959} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.813642] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ea1df9c-69a2-40cf-8f2a-4a765dcb7cf0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.816296] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 880.816510] env[61907]: INFO nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Took 9.34 seconds to spawn the instance on the hypervisor. [ 880.816732] env[61907]: DEBUG nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 880.824165] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad8cf0f8-4b82-426f-aa48-190ef9fe3aaf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.824834] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1243988, 'name': CreateVM_Task, 'duration_secs': 0.420155} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.825718] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 880.827115] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.827115] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.827360] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.829192] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf261b71-6342-432d-95f6-2b8bf44e5fe4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.851056] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 880.851056] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52585c20-4af6-0cee-405d-d72493fc5476" [ 880.851056] env[61907]: _type = "Task" [ 880.851056] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.855830] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52db7a1b-379b-5d0a-98af-f86b60aa8756, 'name': SearchDatastore_Task, 'duration_secs': 0.011552} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.857068] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ea6a169-b8cb-44d7-913e-bb72ddf888d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.865444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.865666] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Processing image c525c73f-cabe-4077-be2c-3ee109d7c7c8 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 880.865906] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.866074] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquired lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.866263] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 880.868714] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ea3391c-b6e8-47aa-9c33-b528735617a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.872750] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 880.872750] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52543b31-b03b-bbc1-741b-c694d3725c12" [ 880.872750] env[61907]: _type = "Task" [ 880.872750] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.878361] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Updated VIF entry in instance network info cache for port d1082c95-320a-4bba-b83c-c3e5da0d2384. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 880.878507] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Updating instance_info_cache with network_info: [{"id": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "address": "fa:16:3e:65:cb:86", "network": {"id": "8ad383b2-b578-47d1-8cc1-7805aae4ee55", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-293583221-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b4cce4c7593147bb8fc26486807edc20", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2e8b74b-aa27-4f31-9414-7bcf531e8642", "external-id": "nsx-vlan-transportzone-544", "segmentation_id": 544, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd1082c95-32", "ovs_interfaceid": "d1082c95-320a-4bba-b83c-c3e5da0d2384", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.883164] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52543b31-b03b-bbc1-741b-c694d3725c12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.895091] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 880.895252] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 880.896246] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6492d509-6d0d-4cd8-b074-de2f09137a35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.902608] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 880.902608] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52570ceb-29cb-4b39-f309-15c687c40394" [ 880.902608] env[61907]: _type = "Task" [ 880.902608] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.914375] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52570ceb-29cb-4b39-f309-15c687c40394, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.047415] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b2f6bb-ca7e-4c2a-ba0d-98f176cb435f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.055150] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0641fec2-475c-417e-9006-192e0c845b84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.085526] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f149b814-fede-4fff-a4bc-2621627a89e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.093143] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a75ccaf6-4edc-4fbb-8b79-3ab1a5b1a151 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.107877] env[61907]: DEBUG nova.compute.provider_tree [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.168015] env[61907]: DEBUG nova.compute.manager [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 881.168850] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00343b0e-e5a5-41b5-a768-6ba869cf3339 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.196232] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.199122] env[61907]: DEBUG nova.compute.manager [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 881.199366] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 881.200151] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d46d1dd-71e0-49a3-b944-6357da3acf77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.207394] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 881.207648] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e40c14c-2958-401a-b438-13e5fa71a713 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.214023] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 881.214023] env[61907]: value = "task-1243989" [ 881.214023] env[61907]: _type = "Task" [ 881.214023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.221677] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243989, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.256576] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-865d386d-c5f0-439b-b50b-b613993ec9b8" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.256831] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 881.257076] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 881.265439] env[61907]: DEBUG nova.compute.manager [req-44322b16-08a4-4a5e-80bb-9ca79e231837 req-bbe07e70-d923-41e3-83a6-3ecb3e0a648d service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Received event network-vif-deleted-aa81555f-9f10-410a-b9ad-12894784fd53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 881.265636] env[61907]: INFO nova.compute.manager [req-44322b16-08a4-4a5e-80bb-9ca79e231837 req-bbe07e70-d923-41e3-83a6-3ecb3e0a648d service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Neutron deleted interface aa81555f-9f10-410a-b9ad-12894784fd53; detaching it from the instance and deleting it from the info cache [ 881.265835] env[61907]: DEBUG nova.network.neutron [req-44322b16-08a4-4a5e-80bb-9ca79e231837 req-bbe07e70-d923-41e3-83a6-3ecb3e0a648d service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.299204] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2f33e39b-5b8b-4a06-923e-8a43185237f1 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.616s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.353016] env[61907]: INFO nova.compute.manager [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Took 38.76 seconds to build instance. [ 881.355964] env[61907]: INFO nova.compute.manager [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Took 36.13 seconds to build instance. [ 881.381651] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52543b31-b03b-bbc1-741b-c694d3725c12, 'name': SearchDatastore_Task, 'duration_secs': 0.015694} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.381949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.382248] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5c357582-0b55-4de7-968e-4daa3522dda4/5c357582-0b55-4de7-968e-4daa3522dda4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 881.382517] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dea9a7ac-d28b-4aa5-b4c2-759f2c13991e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.384990] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Releasing lock "refresh_cache-5c357582-0b55-4de7-968e-4daa3522dda4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.385215] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 881.385434] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.385642] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.385860] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.386061] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] No waiting events found dispatching network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 881.386228] env[61907]: WARNING nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received unexpected event network-vif-plugged-40a79714-91a5-4175-94fe-d7aeae0a8ea5 for instance with vm_state shelved_offloaded and task_state spawning. [ 881.386400] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 881.386597] env[61907]: DEBUG nova.compute.manager [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing instance network info cache due to event network-changed-40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 881.386778] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Acquiring lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.386954] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Acquired lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.387102] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Refreshing network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 881.393414] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 881.393414] env[61907]: value = "task-1243990" [ 881.393414] env[61907]: _type = "Task" [ 881.393414] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.403341] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243990, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.412468] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Preparing fetch location {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 881.412732] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Fetch image to [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793/OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793.vmdk {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 881.412941] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloading stream optimized image c525c73f-cabe-4077-be2c-3ee109d7c7c8 to [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793/OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793.vmdk on the data store datastore2 as vApp {{(pid=61907) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 881.413176] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloading image file data c525c73f-cabe-4077-be2c-3ee109d7c7c8 to the ESX as VM named 'OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793' {{(pid=61907) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 881.468199] env[61907]: DEBUG nova.network.neutron [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.499676] env[61907]: DEBUG nova.objects.instance [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'flavor' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.503204] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 881.503204] env[61907]: value = "resgroup-9" [ 881.503204] env[61907]: _type = "ResourcePool" [ 881.503204] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 881.503506] env[61907]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-21f5c6a1-fcf4-4b75-b09f-728ec47ab0e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.528535] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lease: (returnval){ [ 881.528535] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 881.528535] env[61907]: _type = "HttpNfcLease" [ 881.528535] env[61907]: } obtained for vApp import into resource pool (val){ [ 881.528535] env[61907]: value = "resgroup-9" [ 881.528535] env[61907]: _type = "ResourcePool" [ 881.528535] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 881.528535] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the lease: (returnval){ [ 881.528535] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 881.528535] env[61907]: _type = "HttpNfcLease" [ 881.528535] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 881.536749] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 881.536749] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 881.536749] env[61907]: _type = "HttpNfcLease" [ 881.536749] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 881.610732] env[61907]: DEBUG nova.scheduler.client.report [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 881.646633] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.646965] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.682661] env[61907]: INFO nova.compute.manager [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] instance snapshotting [ 881.685961] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675d394d-e783-4c1c-ad51-e73614469fdc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.715037] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19c8338a-0523-4491-aaa6-e2b4f4ab5347 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.716669] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.727384] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243989, 'name': PowerOffVM_Task, 'duration_secs': 0.360027} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.730595] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 881.730978] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 881.734639] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fee1442b-0db4-46b3-8c35-890ef64138e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.771672] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03229e64-759b-4353-be23-fe5e4ab52fbf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.780976] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea17e0f-e2f9-42b0-9e84-588b63878ee0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.818103] env[61907]: DEBUG nova.compute.manager [req-44322b16-08a4-4a5e-80bb-9ca79e231837 req-bbe07e70-d923-41e3-83a6-3ecb3e0a648d service nova] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Detach interface failed, port_id=aa81555f-9f10-410a-b9ad-12894784fd53, reason: Instance aec01156-431b-447e-88ef-cb8d254c6341 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 881.821022] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 881.821229] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 881.821435] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleting the datastore file [datastore2] 865d386d-c5f0-439b-b50b-b613993ec9b8 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 881.824134] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8689d55a-8747-4308-ac68-9845ccd98137 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.833345] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for the task: (returnval){ [ 881.833345] env[61907]: value = "task-1243993" [ 881.833345] env[61907]: _type = "Task" [ 881.833345] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.847014] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.856096] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5eb5cc73-36ad-49bd-9918-2d4ab78767c8 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.719s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.857693] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b995912e-e98c-4759-ad0f-bbdd49e21233 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.142s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.908794] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243990, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.972312] env[61907]: INFO nova.compute.manager [-] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Took 1.74 seconds to deallocate network for instance. [ 882.006892] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.007083] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.007266] env[61907]: DEBUG nova.network.neutron [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.007464] env[61907]: DEBUG nova.objects.instance [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'info_cache' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.038017] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.038017] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 882.038017] env[61907]: _type = "HttpNfcLease" [ 882.038017] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 882.116665] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.173s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.116895] env[61907]: INFO nova.compute.manager [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Migrating [ 882.117154] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.117309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.118566] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.861s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.121357] env[61907]: INFO nova.compute.claims [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.149459] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 882.190517] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updated VIF entry in instance network info cache for port 40a79714-91a5-4175-94fe-d7aeae0a8ea5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.190775] env[61907]: DEBUG nova.network.neutron [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [{"id": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "address": "fa:16:3e:20:94:16", "network": {"id": "d65c0ae5-526e-4543-9686-66bf55d463ec", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1954411377-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3aab40034a4e4ccdb169abfc11f3b02f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c5f8c7b-74a0-43f7-87b4-3c38dd26ac3e", "external-id": "nsx-vlan-transportzone-988", "segmentation_id": 988, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a79714-91", "ovs_interfaceid": "40a79714-91a5-4175-94fe-d7aeae0a8ea5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.202426] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.238396] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 882.238722] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8c8ac2e3-4a18-4a31-ac46-17d94b870d54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.246599] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 882.246599] env[61907]: value = "task-1243994" [ 882.246599] env[61907]: _type = "Task" [ 882.246599] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.256282] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243994, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.344065] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.408510] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243990, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579628} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.408826] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5c357582-0b55-4de7-968e-4daa3522dda4/5c357582-0b55-4de7-968e-4daa3522dda4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 882.409066] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 882.409331] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fd20f44-6d5c-492f-b1a2-a65e6490b0f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.415849] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 882.415849] env[61907]: value = "task-1243995" [ 882.415849] env[61907]: _type = "Task" [ 882.415849] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.426148] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243995, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.485135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.487218] env[61907]: DEBUG nova.objects.instance [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'flavor' on Instance uuid 41c680fb-5450-43f3-9acb-5218e57a7da9 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.510365] env[61907]: DEBUG nova.objects.base [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Object Instance<92f27804-8974-40c4-9663-b2b72f0bb8e0> lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 882.526453] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.526711] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.526897] env[61907]: DEBUG nova.compute.manager [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.528100] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2344cc31-75c5-443e-b21b-e156f14426ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.537753] env[61907]: DEBUG nova.compute.manager [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 882.538331] env[61907]: DEBUG nova.objects.instance [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'flavor' on Instance uuid 37ff6c54-6b79-4a9c-bc16-29d974185d3e {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.540835] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.540835] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 882.540835] env[61907]: _type = "HttpNfcLease" [ 882.540835] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 882.541302] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 882.541302] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e50c5-cf7f-9fc0-4015-2b87fc716497" [ 882.541302] env[61907]: _type = "HttpNfcLease" [ 882.541302] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 882.541989] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b05e0e5b-a99f-450c-ba18-0b99853fd1a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.549766] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 882.550076] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk. {{(pid=61907) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 882.615322] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-70860206-ebae-4918-9f8d-bbf1a4ab8c31 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.624364] env[61907]: INFO nova.compute.rpcapi [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Automatically selected compute RPC version 6.4 from minimum service version 68 [ 882.624883] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.673208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.697352] env[61907]: DEBUG oslo_concurrency.lockutils [req-cb246bb8-cb2b-4842-b9d4-cb4868980272 req-167ba8f9-8976-46c5-9724-7d49754adba9 service nova] Releasing lock "refresh_cache-691fff4a-9dbd-47f2-8908-69057a2bb6fe" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.699163] env[61907]: DEBUG nova.objects.instance [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lazy-loading 'flavor' on Instance uuid da490943-511b-4776-8f16-4f51c3b055c4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.706558] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.757258] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243994, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.831802] env[61907]: DEBUG nova.compute.manager [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 882.832904] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac15577e-b47a-44d5-b45b-4acaa1d4418b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.848436] env[61907]: DEBUG oslo_vmware.api [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Task: {'id': task-1243993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.60419} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.848436] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 882.848650] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 882.848837] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 882.849195] env[61907]: INFO nova.compute.manager [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Took 1.65 seconds to destroy the instance on the hypervisor. [ 882.849490] env[61907]: DEBUG oslo.service.loopingcall [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.849700] env[61907]: DEBUG nova.compute.manager [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 882.849798] env[61907]: DEBUG nova.network.neutron [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 882.925309] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243995, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074328} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.925580] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.926469] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da53cfb-f892-49ab-929d-47cfabe17899 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.952315] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] 5c357582-0b55-4de7-968e-4daa3522dda4/5c357582-0b55-4de7-968e-4daa3522dda4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.952651] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2a18ee2f-24b3-4f53-98d0-b71ec9c6ca79 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.972088] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 882.972088] env[61907]: value = "task-1243996" [ 882.972088] env[61907]: _type = "Task" [ 882.972088] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.979755] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.993386] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.993386] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquired lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.993622] env[61907]: DEBUG nova.network.neutron [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.993713] env[61907]: DEBUG nova.objects.instance [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'info_cache' on Instance uuid 41c680fb-5450-43f3-9acb-5218e57a7da9 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.148939] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.148939] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.148939] env[61907]: DEBUG nova.network.neutron [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 883.209822] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.210754] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.210754] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.261741] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243994, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.301381] env[61907]: DEBUG nova.compute.manager [req-243fcf56-ea26-4322-9b57-57723c6c8592 req-57f25421-6d8e-4348-922b-98ad4ad585ca service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Received event network-vif-deleted-05a00a60-c277-47b8-98e8-bc8f0f46258a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 883.302015] env[61907]: INFO nova.compute.manager [req-243fcf56-ea26-4322-9b57-57723c6c8592 req-57f25421-6d8e-4348-922b-98ad4ad585ca service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Neutron deleted interface 05a00a60-c277-47b8-98e8-bc8f0f46258a; detaching it from the instance and deleting it from the info cache [ 883.302015] env[61907]: DEBUG nova.network.neutron [req-243fcf56-ea26-4322-9b57-57723c6c8592 req-57f25421-6d8e-4348-922b-98ad4ad585ca service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.350918] env[61907]: INFO nova.compute.manager [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] instance snapshotting [ 883.357197] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53110484-cb99-4cd7-89ba-1547de1b160d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.395016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032ca649-d4db-4262-b3f1-ba98c3028805 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.499868] env[61907]: DEBUG nova.objects.base [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Object Instance<41c680fb-5450-43f3-9acb-5218e57a7da9> lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 883.501646] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243996, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.502224] env[61907]: DEBUG nova.network.neutron [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.524305] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-311868fd-3c88-4f27-83e6-16476adddb5a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.536451] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea87503-251e-4af4-94b7-25bff1a71818 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.572653] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.580414] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-417fe813-7205-49f6-91e3-ea81c18049fc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.583239] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e7fcc37-77ae-4203-9f31-5e87e85684e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.595147] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5f4fb6-59d3-4fd6-9104-134c658ed098 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.599579] env[61907]: DEBUG oslo_vmware.api [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 883.599579] env[61907]: value = "task-1243997" [ 883.599579] env[61907]: _type = "Task" [ 883.599579] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.615482] env[61907]: DEBUG nova.network.neutron [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.617200] env[61907]: DEBUG nova.compute.provider_tree [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.627630] env[61907]: DEBUG oslo_vmware.api [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243997, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.703458] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.761544] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1243994, 'name': CreateSnapshot_Task, 'duration_secs': 1.186671} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.763941] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 883.764845] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-286ad0aa-1fb8-4826-a00e-0c9913ff7abb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.807806] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f6b82fa9-16f6-4641-9b49-3a2834fc5957 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.827641] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa91b6c-1190-4011-a5d2-0f331b20ee6c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.870856] env[61907]: DEBUG nova.compute.manager [req-243fcf56-ea26-4322-9b57-57723c6c8592 req-57f25421-6d8e-4348-922b-98ad4ad585ca service nova] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Detach interface failed, port_id=05a00a60-c277-47b8-98e8-bc8f0f46258a, reason: Instance 865d386d-c5f0-439b-b50b-b613993ec9b8 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 883.917445] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 883.917787] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1d7ea9c5-0d29-4588-8746-d8e10d40d660 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.926145] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 883.926145] env[61907]: value = "task-1243998" [ 883.926145] env[61907]: _type = "Task" [ 883.926145] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.940524] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243998, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.986363] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243996, 'name': ReconfigVM_Task, 'duration_secs': 0.689133} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.988835] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Reconfigured VM instance instance-0000004d to attach disk [datastore2] 5c357582-0b55-4de7-968e-4daa3522dda4/5c357582-0b55-4de7-968e-4daa3522dda4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.988835] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd343563-f8ca-459f-b400-26523f5267bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.996023] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 883.996023] env[61907]: value = "task-1243999" [ 883.996023] env[61907]: _type = "Task" [ 883.996023] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.010960] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.012563] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243999, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.115362] env[61907]: DEBUG oslo_vmware.api [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1243997, 'name': PowerOffVM_Task, 'duration_secs': 0.180587} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.115642] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 884.115848] env[61907]: DEBUG nova.compute.manager [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 884.116789] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fed502-192f-4059-aab0-aba90a81a6d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.124221] env[61907]: DEBUG nova.scheduler.client.report [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 884.127652] env[61907]: INFO nova.compute.manager [-] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Took 1.28 seconds to deallocate network for instance. [ 884.205238] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.286069] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 884.287604] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3d34ca0c-8318-4ee0-b785-c2406af2458d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.302010] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 884.302010] env[61907]: value = "task-1244000" [ 884.302010] env[61907]: _type = "Task" [ 884.302010] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.310203] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244000, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.325207] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Completed reading data from the image iterator. {{(pid=61907) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 884.325705] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 884.326353] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06337640-4c83-4966-a4d5-791fcebefb39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.332901] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 884.333233] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 884.334536] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-e51e0b6b-02bc-450b-81f9-0b074bb84008 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.418245] env[61907]: DEBUG nova.network.neutron [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 884.436405] env[61907]: DEBUG nova.network.neutron [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.448426] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243998, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.510564] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1243999, 'name': Rename_Task, 'duration_secs': 0.153078} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.511426] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.511710] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a445c908-872f-4ccd-837e-c4d629a8c492 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.524685] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 884.524685] env[61907]: value = "task-1244001" [ 884.524685] env[61907]: _type = "Task" [ 884.524685] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.534081] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.635622] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.636439] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 884.640480] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.155s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.640667] env[61907]: DEBUG nova.objects.instance [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lazy-loading 'resources' on Instance uuid aec01156-431b-447e-88ef-cb8d254c6341 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.644683] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.645935] env[61907]: DEBUG nova.network.neutron [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Updating instance_info_cache with network_info: [{"id": "4d440707-544e-46dc-bf63-b4910c232a7a", "address": "fa:16:3e:15:8c:a3", "network": {"id": "eb0b7618-0ba6-4515-bea6-0b741c3d77ff", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-412258735-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87439b1925b644888b9ed1a52ddff936", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "822050c7-1845-485d-b87e-73778d21c33c", "external-id": "nsx-vlan-transportzone-701", "segmentation_id": 701, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4d440707-54", "ovs_interfaceid": "4d440707-544e-46dc-bf63-b4910c232a7a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.647922] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a03260-3ca6-49d5-bdf9-7e376e66c7f2 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.121s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.704493] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.730696] env[61907]: DEBUG oslo_vmware.rw_handles [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52082d22-ee07-66a9-b1bd-5b875295989b/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 884.730696] env[61907]: INFO nova.virt.vmwareapi.images [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Downloaded image file data c525c73f-cabe-4077-be2c-3ee109d7c7c8 [ 884.731229] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e4192b-5ab1-4a97-8801-92e3e256b890 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.747577] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42a5d906-33d7-4081-be3e-ed4be38a6465 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.773981] env[61907]: INFO nova.virt.vmwareapi.images [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] The imported VM was unregistered [ 884.776304] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Caching image {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 884.776542] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Creating directory with path [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.777177] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5c4e1d78-9e3a-43da-9140-4245f073c99d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.788039] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Created directory with path [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.788251] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793/OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793.vmdk to [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk. {{(pid=61907) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 884.788537] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-ec92eec7-9032-4d5b-80f7-6400320eab3c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.795054] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 884.795054] env[61907]: value = "task-1244003" [ 884.795054] env[61907]: _type = "Task" [ 884.795054] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.804272] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.812493] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244000, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.937626] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1243998, 'name': CreateSnapshot_Task, 'duration_secs': 0.737116} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.937893] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 884.938903] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24578cd5-2d79-491e-9707-fabeaaf5e823 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.942873] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.012174] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.012488] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.012719] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.012915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.013209] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.016167] env[61907]: INFO nova.compute.manager [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Terminating instance [ 885.022826] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 885.023224] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e54eccb4-8b84-458d-ab06-c6c8a33b1cba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.042449] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244001, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.042824] env[61907]: DEBUG oslo_vmware.api [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 885.042824] env[61907]: value = "task-1244004" [ 885.042824] env[61907]: _type = "Task" [ 885.042824] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.144634] env[61907]: DEBUG nova.compute.utils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.150036] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 885.150036] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.154228] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Releasing lock "refresh_cache-41c680fb-5450-43f3-9acb-5218e57a7da9" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.204876] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.271521] env[61907]: DEBUG nova.policy [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7e8057170d8a41d2a49bf4485c352e3b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '89e4f90a5fe44853a926ceba2f5150dd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.317796] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.334211] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244000, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.376350] env[61907]: DEBUG nova.compute.manager [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 885.376513] env[61907]: DEBUG nova.compute.manager [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing instance network info cache due to event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 885.376721] env[61907]: DEBUG oslo_concurrency.lockutils [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.478193] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 885.482666] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-249da9c3-7948-41ef-b824-4d9fd235afff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.489997] env[61907]: DEBUG nova.network.neutron [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.494056] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 885.494056] env[61907]: value = "task-1244005" [ 885.494056] env[61907]: _type = "Task" [ 885.494056] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.505205] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244005, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.526031] env[61907]: DEBUG nova.compute.manager [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 885.526031] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 885.526031] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246be5cc-6a0f-4323-ba85-bfd251a3fed3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.534059] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f376e5-f8dc-4735-87e8-92789c0f3eed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.539490] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 885.540296] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de94c2d5-5635-44ea-99c7-dbfe150e2c18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.545439] env[61907]: DEBUG oslo_vmware.api [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244001, 'name': PowerOnVM_Task, 'duration_secs': 0.705024} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.549078] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.549503] env[61907]: INFO nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Took 10.63 seconds to spawn the instance on the hypervisor. [ 885.549911] env[61907]: DEBUG nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 885.552648] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282e9eeb-ae27-4e98-a14d-1c468caf486a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.557357] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d8fbf0-c833-48bd-b0cb-e9cab3aef214 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.564477] env[61907]: DEBUG oslo_vmware.api [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244004, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.600715] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3386d87f-6c05-4190-b9e1-19cd9c7edff0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.618551] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffdf16b-ce38-4e3b-9e25-9abc598113c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.639185] env[61907]: DEBUG nova.compute.provider_tree [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.654424] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 885.684301] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 885.684583] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 885.684729] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleting the datastore file [datastore2] 37ff6c54-6b79-4a9c-bc16-29d974185d3e {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.685035] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4cb0f965-f6da-4d6a-9ae0-c7a745258fb1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.696028] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 885.696028] env[61907]: value = "task-1244007" [ 885.696028] env[61907]: _type = "Task" [ 885.696028] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.710651] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.715328] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.747618] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Successfully created port: 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.817056] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.821292] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244000, 'name': CloneVM_Task} progress is 95%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.993338] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.993698] env[61907]: DEBUG nova.compute.manager [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Inject network info {{(pid=61907) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7562}} [ 885.994149] env[61907]: DEBUG nova.compute.manager [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] network_info to inject: |[{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7563}} [ 885.999738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfiguring VM instance to set the machine id {{(pid=61907) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 886.000224] env[61907]: DEBUG oslo_concurrency.lockutils [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.000508] env[61907]: DEBUG nova.network.neutron [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 886.001969] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a5c50cf-7ec2-44bb-b408-7dd5c94205e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.032031] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244005, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.033736] env[61907]: DEBUG oslo_vmware.api [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 886.033736] env[61907]: value = "task-1244008" [ 886.033736] env[61907]: _type = "Task" [ 886.033736] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.046208] env[61907]: DEBUG oslo_vmware.api [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244008, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.060580] env[61907]: DEBUG oslo_vmware.api [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244004, 'name': PowerOnVM_Task, 'duration_secs': 0.855857} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.060775] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.060964] env[61907]: DEBUG nova.compute.manager [None req-4e81543a-be70-49ad-a34b-015df82d03ee tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 886.061876] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b56b72c-7281-4f82-95da-c1f9ed3717a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.120904] env[61907]: INFO nova.compute.manager [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Took 34.00 seconds to build instance. [ 886.143262] env[61907]: DEBUG nova.scheduler.client.report [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 886.163940] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.164285] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-512d4ac6-39f4-41b9-8b35-da23281b5524 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.174186] env[61907]: DEBUG oslo_vmware.api [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 886.174186] env[61907]: value = "task-1244009" [ 886.174186] env[61907]: _type = "Task" [ 886.174186] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.185417] env[61907]: DEBUG oslo_vmware.api [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244009, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.215924] env[61907]: DEBUG oslo_vmware.api [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1243984, 'name': ReconfigVM_Task, 'duration_secs': 5.814037} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.222549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.223184] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Reconfigured VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 886.226386] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.316577] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.327151] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244000, 'name': CloneVM_Task, 'duration_secs': 1.763597} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.327853] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Created linked-clone VM from snapshot [ 886.329374] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5349523-3b95-48fc-bed0-2610ea194003 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.343551] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Uploading image 498f7820-6901-4f82-93d8-38e23511893d {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 886.358704] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 886.359086] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-554bc1c9-4705-4e8d-bc7b-3ed8bdec9a41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.373034] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 886.373034] env[61907]: value = "task-1244010" [ 886.373034] env[61907]: _type = "Task" [ 886.373034] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.388178] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244010, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.500032] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6599ed-1dcd-4ec7-b803-3eeef71062d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.508606] env[61907]: DEBUG nova.objects.instance [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lazy-loading 'flavor' on Instance uuid da490943-511b-4776-8f16-4f51c3b055c4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.532047] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 886.545913] env[61907]: DEBUG oslo_concurrency.lockutils [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.547845] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244005, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.561769] env[61907]: DEBUG oslo_vmware.api [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244008, 'name': ReconfigVM_Task, 'duration_secs': 0.207153} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.562418] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a654f1e5-1265-42c5-91a4-d2825eb59d36 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfigured VM instance to set the machine id {{(pid=61907) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 886.626151] env[61907]: DEBUG oslo_concurrency.lockutils [None req-40b48487-5a8a-4a57-b96f-883f8124dd1c tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.567s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.648373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.008s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.651552] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.978s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.656026] env[61907]: INFO nova.compute.claims [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 886.666049] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 886.677033] env[61907]: INFO nova.scheduler.client.report [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Deleted allocations for instance aec01156-431b-447e-88ef-cb8d254c6341 [ 886.701295] env[61907]: DEBUG oslo_vmware.api [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244009, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.704878] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.705351] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.705698] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.706842] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.707181] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.707478] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.708228] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.708467] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.708810] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.709220] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.709543] env[61907]: DEBUG nova.virt.hardware [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.715522] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9779e29-2b94-44d4-90ac-d31b8a762449 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.728052] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.733538] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f24399-a8eb-43d0-ab83-15bd482d674a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.815633] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.870212] env[61907]: DEBUG nova.network.neutron [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updated VIF entry in instance network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.870808] env[61907]: DEBUG nova.network.neutron [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.889525] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244010, 'name': Destroy_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.020331] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244005, 'name': CloneVM_Task} progress is 95%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.050931] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.051602] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-24c63f03-cee8-428e-8ece-00cc0a7707df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.064642] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 887.064642] env[61907]: value = "task-1244011" [ 887.064642] env[61907]: _type = "Task" [ 887.064642] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.085670] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244011, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.191024] env[61907]: DEBUG oslo_vmware.api [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244009, 'name': PowerOnVM_Task, 'duration_secs': 0.71301} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.194168] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.194747] env[61907]: DEBUG nova.compute.manager [None req-32e50912-7dca-4c07-b338-87bfdc0e5418 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 887.195788] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07fbf35e-8b42-494a-9d34-afdaf47eb91b tempest-ServersNegativeTestMultiTenantJSON-1313387767 tempest-ServersNegativeTestMultiTenantJSON-1313387767-project-member] Lock "aec01156-431b-447e-88ef-cb8d254c6341" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.357s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.199105] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed3cec1-a9a2-4254-8b51-e623ed125a28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.219768] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244007, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.308798] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.333557] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "5c357582-0b55-4de7-968e-4daa3522dda4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.334289] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.335056] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.335056] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.335270] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.338024] env[61907]: INFO nova.compute.manager [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Terminating instance [ 887.374976] env[61907]: DEBUG oslo_concurrency.lockutils [req-549433be-6fa4-4077-8e53-270f7ba29a35 req-620f4def-b215-4aad-bfd6-ab2c272e4fc3 service nova] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.375620] env[61907]: DEBUG oslo_concurrency.lockutils [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.387067] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244010, 'name': Destroy_Task, 'duration_secs': 0.799458} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.387362] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Destroyed the VM [ 887.387681] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 887.387941] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-d4e59fb8-4a60-4bed-8805-e56217aff103 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.394344] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 887.394344] env[61907]: value = "task-1244012" [ 887.394344] env[61907]: _type = "Task" [ 887.394344] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.402577] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244012, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.454644] env[61907]: DEBUG nova.compute.manager [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-deleted-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 887.454644] env[61907]: INFO nova.compute.manager [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Neutron deleted interface 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed; detaching it from the instance and deleting it from the info cache [ 887.454987] env[61907]: DEBUG nova.network.neutron [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "address": "fa:16:3e:ea:e2:90", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0ff98c3-b1", "ovs_interfaceid": "b0ff98c3-b17f-497e-8ba2-0578ebbd2927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.514054] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244005, 'name': CloneVM_Task, 'duration_secs': 1.866802} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.514389] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Created linked-clone VM from snapshot [ 887.515139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80737a28-14c7-4170-a3b5-2e44ed3d1f71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.523770] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Uploading image 93afda2d-a435-4b9d-983e-30d2ad798313 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 887.557559] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.557559] env[61907]: value = "vm-268275" [ 887.557559] env[61907]: _type = "VirtualMachine" [ 887.557559] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.557559] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7d85c267-3029-4248-985e-a8dd43e5b33b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.568909] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease: (returnval){ [ 887.568909] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520eaf8b-5780-1e69-8e1f-b466ee206c40" [ 887.568909] env[61907]: _type = "HttpNfcLease" [ 887.568909] env[61907]: } obtained for exporting VM: (result){ [ 887.568909] env[61907]: value = "vm-268275" [ 887.568909] env[61907]: _type = "VirtualMachine" [ 887.568909] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.568909] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the lease: (returnval){ [ 887.568909] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520eaf8b-5780-1e69-8e1f-b466ee206c40" [ 887.568909] env[61907]: _type = "HttpNfcLease" [ 887.568909] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.579329] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244011, 'name': PowerOffVM_Task, 'duration_secs': 0.31274} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.581319] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 887.582285] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 887.584949] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.584949] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520eaf8b-5780-1e69-8e1f-b466ee206c40" [ 887.584949] env[61907]: _type = "HttpNfcLease" [ 887.584949] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.716052] env[61907]: DEBUG oslo_vmware.api [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244007, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.610645} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.716748] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.716974] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 887.717317] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 887.717354] env[61907]: INFO nova.compute.manager [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Took 2.19 seconds to destroy the instance on the hypervisor. [ 887.717598] env[61907]: DEBUG oslo.service.loopingcall [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.717798] env[61907]: DEBUG nova.compute.manager [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 887.717949] env[61907]: DEBUG nova.network.neutron [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 887.817403] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244003, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.548318} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.817721] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793/OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793.vmdk to [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk. [ 887.817912] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Cleaning up location [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 887.818107] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_edbeaf6d-cfd8-45d1-af5a-4551f87e3793 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 887.818472] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f923d0e4-ea21-4233-b31f-e9bd5b2308ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.827526] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 887.827526] env[61907]: value = "task-1244014" [ 887.827526] env[61907]: _type = "Task" [ 887.827526] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.838066] env[61907]: DEBUG nova.compute.manager [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-deleted-b0ff98c3-b17f-497e-8ba2-0578ebbd2927 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 887.838362] env[61907]: INFO nova.compute.manager [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Neutron deleted interface b0ff98c3-b17f-497e-8ba2-0578ebbd2927; detaching it from the instance and deleting it from the info cache [ 887.838639] env[61907]: DEBUG nova.network.neutron [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.844120] env[61907]: DEBUG nova.compute.manager [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 887.844369] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 887.844929] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.845928] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2819517-0a8d-4f12-aeeb-bdfee85fbe83 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.855268] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 887.855538] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-925a1d8a-bc78-4e06-a0f2-672b1a5627c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.864324] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 887.864324] env[61907]: value = "task-1244015" [ 887.864324] env[61907]: _type = "Task" [ 887.864324] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.884392] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244015, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.911457] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244012, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.960690] env[61907]: DEBUG oslo_concurrency.lockutils [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.960690] env[61907]: DEBUG oslo_concurrency.lockutils [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] Acquired lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.960690] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4297584f-3113-49c8-b012-45b8237229b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.978074] env[61907]: DEBUG oslo_concurrency.lockutils [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] Releasing lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.978580] env[61907]: WARNING nova.compute.manager [req-aa598567-fa44-4912-a711-f276df281d9d req-32922722-4408-4ef4-9d29-975a5b5b6c33 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Detach interface failed, port_id=0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed, reason: No device with interface-id 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed exists on VM: nova.exception.NotFound: No device with interface-id 0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed exists on VM [ 888.050095] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.050445] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.050845] env[61907]: DEBUG nova.network.neutron [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.053577] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9bc15f-d44e-417d-b95d-1ddb993bfab7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.065725] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba55a0c-c452-4f04-b421-d840e438be65 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.082508] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.082508] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520eaf8b-5780-1e69-8e1f-b466ee206c40" [ 888.082508] env[61907]: _type = "HttpNfcLease" [ 888.082508] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 888.133789] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 888.133789] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520eaf8b-5780-1e69-8e1f-b466ee206c40" [ 888.133789] env[61907]: _type = "HttpNfcLease" [ 888.133789] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 888.137077] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.137511] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.137812] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.138125] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.138402] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.138678] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.139113] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.139362] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.139669] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.139873] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.140091] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.150840] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25917331-c30a-44ee-a967-bb8e56290ee8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.155126] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b802316-89f9-4eaf-bd0b-a41804c406d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.175278] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12af33c-204e-4928-8833-0534d403b49e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.180893] env[61907]: DEBUG nova.network.neutron [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.193468] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed60368c-cf14-4bac-aa37-5933440452b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.198058] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 888.198260] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 888.201700] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 888.201700] env[61907]: value = "task-1244016" [ 888.201700] env[61907]: _type = "Task" [ 888.201700] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.275685] env[61907]: DEBUG nova.compute.provider_tree [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.283827] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244016, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.303493] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.303788] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.304084] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.304279] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.304488] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.309127] env[61907]: INFO nova.compute.manager [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Terminating instance [ 888.339289] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042731} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.339640] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.339967] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Releasing lock "[datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.340093] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk to [datastore2] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 888.340360] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d9d7bf0-3b77-4eeb-ad30-1b28ec7cd00c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.348364] env[61907]: DEBUG oslo_concurrency.lockutils [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] Acquiring lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.350490] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-bad7cbb3-eef1-4b25-b8bf-3d2ac50cad3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.352990] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 888.352990] env[61907]: value = "task-1244017" [ 888.352990] env[61907]: _type = "Task" [ 888.352990] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.365793] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.377608] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244015, 'name': PowerOffVM_Task, 'duration_secs': 0.447234} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.377608] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 888.377608] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 888.379396] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-16a103c8-6ba9-4fd8-8c7b-eb54e65c765f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.382490] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Successfully updated port: 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 888.410548] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244012, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.459421] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 888.459681] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 888.459923] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Deleting the datastore file [datastore2] 5c357582-0b55-4de7-968e-4daa3522dda4 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.460197] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e7d3782e-eb22-461e-8560-4aa3b97dceae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.469489] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for the task: (returnval){ [ 888.469489] env[61907]: value = "task-1244019" [ 888.469489] env[61907]: _type = "Task" [ 888.469489] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.488888] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.721692] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244016, 'name': ReconfigVM_Task, 'duration_secs': 0.394758} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.723712] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 888.779409] env[61907]: DEBUG nova.scheduler.client.report [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 888.811264] env[61907]: DEBUG nova.compute.manager [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 888.811530] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 888.815904] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68be218c-749f-42b6-839e-8746aaff00cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.829558] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 888.830185] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d2e69e4-8b6b-4855-9afa-56c4c90af1da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.840666] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 888.840666] env[61907]: value = "task-1244020" [ 888.840666] env[61907]: _type = "Task" [ 888.840666] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.853865] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.870420] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.884882] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.885077] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.885395] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.917942] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244012, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.956306] env[61907]: DEBUG nova.network.neutron [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.984496] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.191857] env[61907]: DEBUG nova.network.neutron [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [{"id": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "address": "fa:16:3e:c5:65:d7", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.142", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce4e1e2e-36", "ovs_interfaceid": "ce4e1e2e-3671-4aba-a760-b96db6f21180", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.233620] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.234165] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.234796] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.234796] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.234904] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.235050] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.235287] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.235458] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.235630] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.235901] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.236699] env[61907]: DEBUG nova.virt.hardware [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.245107] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 889.246373] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2c5b5ba-3015-4b82-bf1d-57a2f6662bd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.267625] env[61907]: DEBUG nova.network.neutron [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.274706] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 889.274706] env[61907]: value = "task-1244021" [ 889.274706] env[61907]: _type = "Task" [ 889.274706] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.286481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.635s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.287388] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 889.296907] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.652s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.297608] env[61907]: DEBUG nova.objects.instance [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lazy-loading 'resources' on Instance uuid 865d386d-c5f0-439b-b50b-b613993ec9b8 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.300181] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244021, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.366640] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244020, 'name': PowerOffVM_Task, 'duration_secs': 0.263337} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.368275] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 889.368996] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 889.370643] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-177c314f-7c0d-4272-8318-11a62d547a62 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.380981] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.423997] env[61907]: DEBUG oslo_vmware.api [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244012, 'name': RemoveSnapshot_Task, 'duration_secs': 1.83854} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.424464] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 889.459297] env[61907]: INFO nova.compute.manager [-] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Took 1.74 seconds to deallocate network for instance. [ 889.468698] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 889.489330] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244019, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.692080] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-vif-plugged-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 889.692408] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.692711] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.692920] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.693212] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] No waiting events found dispatching network-vif-plugged-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 889.693454] env[61907]: WARNING nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received unexpected event network-vif-plugged-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 for instance with vm_state building and task_state spawning. [ 889.693652] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 889.693879] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing instance network info cache due to event network-changed-e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 889.694208] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Acquiring lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.699030] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.757697] env[61907]: DEBUG nova.network.neutron [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.770617] env[61907]: DEBUG oslo_concurrency.lockutils [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.771177] env[61907]: DEBUG nova.compute.manager [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Inject network info {{(pid=61907) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7562}} [ 889.772075] env[61907]: DEBUG nova.compute.manager [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] network_info to inject: |[{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7563}} [ 889.781705] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfiguring VM instance to set the machine id {{(pid=61907) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1796}} [ 889.783051] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Acquired lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.783115] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Refreshing network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.784807] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88c53e19-868a-41f3-9967-af29d1ff8f28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.814538] env[61907]: DEBUG nova.compute.utils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 889.824165] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 889.824554] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.841202] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244021, 'name': ReconfigVM_Task, 'duration_secs': 0.264465} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.843499] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 889.844455] env[61907]: DEBUG oslo_vmware.api [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 889.844455] env[61907]: value = "task-1244023" [ 889.844455] env[61907]: _type = "Task" [ 889.844455] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.846074] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51eef18-1c1d-48c2-8364-23a9e79375b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.890090] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.895181] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9e14b95-468b-4a3e-b902-918fefbd81aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.910731] env[61907]: DEBUG oslo_vmware.api [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244023, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.920353] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.929355] env[61907]: WARNING nova.compute.manager [None req-b7235cdb-2d1a-4399-bc92-201f671f3985 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Image not found during snapshot: nova.exception.ImageNotFound: Image 498f7820-6901-4f82-93d8-38e23511893d could not be found. [ 889.934061] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 889.934061] env[61907]: value = "task-1244024" [ 889.934061] env[61907]: _type = "Task" [ 889.934061] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.946956] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244024, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.969505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.973426] env[61907]: DEBUG nova.policy [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '346a1c2a6b5a4ed5aa93c9c13a1610a9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0b2c5e8cc184e328a19e0c0a8cf8169', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 889.989017] env[61907]: DEBUG oslo_vmware.api [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Task: {'id': task-1244019, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.128219} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.989432] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.989682] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 889.989922] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 889.990196] env[61907]: INFO nova.compute.manager [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Took 2.15 seconds to destroy the instance on the hypervisor. [ 889.990498] env[61907]: DEBUG oslo.service.loopingcall [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.990748] env[61907]: DEBUG nova.compute.manager [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 889.990869] env[61907]: DEBUG nova.network.neutron [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.200114] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6820d8-057c-4d70-8b9b-224ae42cc10c tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-52404147-0375-4f9c-9e1e-4a52001df7cb-0c78d5e6-ccdc-42b6-834a-96a6f8aa66ed" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.455s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.214790] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 890.215440] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 890.217082] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleting the datastore file [datastore1] 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.218115] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28a34a12-3095-438f-b617-3fd3d5aabf5c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.236934] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 890.236934] env[61907]: value = "task-1244025" [ 890.236934] env[61907]: _type = "Task" [ 890.236934] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.249780] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b690fe0-1a57-4d04-acae-e39285b0fae1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.262267] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.262844] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.263357] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Instance network_info: |[{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 890.264361] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:b4:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd19577c9-1b2e-490b-8031-2f278dd3f570', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 890.272315] env[61907]: DEBUG oslo.service.loopingcall [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.274846] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 890.275272] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac39e593-47e3-44fd-be97-eb23194e07a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.291246] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a36cf2-4008-4483-b640-f1d166ce7b1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.302288] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 890.302288] env[61907]: value = "task-1244026" [ 890.302288] env[61907]: _type = "Task" [ 890.302288] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.329386] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 890.337067] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f5b799-e03e-4575-b55b-07c47d15c1f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.353498] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5d50e7-65e3-49bf-84d8-56ff53079450 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.357704] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244026, 'name': CreateVM_Task} progress is 15%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.371596] env[61907]: DEBUG oslo_vmware.api [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244023, 'name': ReconfigVM_Task, 'duration_secs': 0.258241} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.384322] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-466528cb-a630-40f2-a039-3d3657b92eba tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Reconfigured VM instance to set the machine id {{(pid=61907) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1799}} [ 890.387516] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.388090] env[61907]: DEBUG nova.compute.provider_tree [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.452052] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244024, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.757161] env[61907]: DEBUG oslo_vmware.api [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3315} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.758365] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.758681] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 890.758887] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 890.759237] env[61907]: INFO nova.compute.manager [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Took 1.95 seconds to destroy the instance on the hypervisor. [ 890.759599] env[61907]: DEBUG oslo.service.loopingcall [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 890.760500] env[61907]: DEBUG nova.compute.manager [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 890.760665] env[61907]: DEBUG nova.network.neutron [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 890.843304] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244026, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.845448] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updated VIF entry in instance network info cache for port e34b5242-a01d-49e3-9aba-aaf54bf279df. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.845930] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [{"id": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "address": "fa:16:3e:45:7b:c2", "network": {"id": "e3a5b197-81ae-4de6-8c0e-67a8adb55e26", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1670781230-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e704030091a4d7985062c2a39d4f1b5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e59b364d-b7f6-499d-b7dc-82b8a819aa12", "external-id": "nsx-vlan-transportzone-731", "segmentation_id": 731, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape34b5242-a0", "ovs_interfaceid": "e34b5242-a01d-49e3-9aba-aaf54bf279df", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.872607] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.894338] env[61907]: DEBUG nova.scheduler.client.report [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 890.949035] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244024, 'name': ReconfigVM_Task, 'duration_secs': 0.559536} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.949699] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Reconfigured VM instance instance-00000048 to attach disk [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3/0bfdb2d0-8388-4be8-a2ee-743c029db6c3.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.950228] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 890.960144] env[61907]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port b0ff98c3-b17f-497e-8ba2-0578ebbd2927 could not be found.", "detail": ""}} {{(pid=61907) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 890.960144] env[61907]: DEBUG nova.network.neutron [-] Unable to show port b0ff98c3-b17f-497e-8ba2-0578ebbd2927 as it no longer exists. {{(pid=61907) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 891.080595] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.080687] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.155697] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Successfully created port: 5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 891.241408] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.241703] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.241950] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.242165] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.242353] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.244872] env[61907]: INFO nova.compute.manager [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Terminating instance [ 891.341971] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 891.343689] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244026, 'name': CreateVM_Task, 'duration_secs': 0.572033} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.344564] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.345503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.345660] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.346448] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 891.346745] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0cf55d57-d0ae-482a-975b-9e6f658992d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.348919] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Releasing lock "refresh_cache-da490943-511b-4776-8f16-4f51c3b055c4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.349184] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 891.349364] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing instance network info cache due to event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 891.349603] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.350347] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.350347] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.356103] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 891.356103] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e96871-a36a-28ee-a7aa-619a58f24795" [ 891.356103] env[61907]: _type = "Task" [ 891.356103] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.376294] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e96871-a36a-28ee-a7aa-619a58f24795, 'name': SearchDatastore_Task, 'duration_secs': 0.010888} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.378881] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.379188] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.379432] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.379887] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.379887] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.380045] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.380206] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.380373] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.380559] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.380799] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.381141] env[61907]: DEBUG nova.virt.hardware [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.384921] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.385114] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.385383] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.385565] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.385837] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.387060] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a63615-ebe7-471c-884b-f95355d58532 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.389869] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244017, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.888623} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.390119] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-273f6153-7e70-42a7-9415-7f207d742204 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.391922] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/c525c73f-cabe-4077-be2c-3ee109d7c7c8/c525c73f-cabe-4077-be2c-3ee109d7c7c8.vmdk to [datastore2] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 891.393586] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43664e54-d00a-4758-b8c9-20c92ed19b5c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.399331] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.103s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.402474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0f106b-252b-488b-b59e-4e32b82e66b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.416747] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.447s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.417016] env[61907]: DEBUG nova.objects.instance [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'resources' on Instance uuid 37ff6c54-6b79-4a9c-bc16-29d974185d3e {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.427267] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.427656] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.427845] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.432200] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c4f6cc28-e30d-4c30-a0d7-92d117737dad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.447361] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad4ee411-49fa-4e31-8e10-d57aa1039353 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.458534] env[61907]: INFO nova.scheduler.client.report [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Deleted allocations for instance 865d386d-c5f0-439b-b50b-b613993ec9b8 [ 891.464836] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cf86b3-0b8b-4eef-a682-decda88f2b1c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.471958] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 891.471958] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4186e-3ac7-426b-48fc-6afcce7eb8e3" [ 891.471958] env[61907]: _type = "Task" [ 891.471958] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.476188] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 891.476188] env[61907]: value = "task-1244027" [ 891.476188] env[61907]: _type = "Task" [ 891.476188] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.499138] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea481ae0-8327-4221-88d8-c56989615768 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.505908] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4186e-3ac7-426b-48fc-6afcce7eb8e3, 'name': SearchDatastore_Task, 'duration_secs': 0.012618} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.507361] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19e2b509-4f40-437c-9a50-e83d958de70d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.529483] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 891.533304] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.541029] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 891.541029] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5ac6a-45de-4f93-f70d-15aef25f11d8" [ 891.541029] env[61907]: _type = "Task" [ 891.541029] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.551166] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5ac6a-45de-4f93-f70d-15aef25f11d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.586305] env[61907]: DEBUG nova.compute.utils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 891.607518] env[61907]: DEBUG nova.network.neutron [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.749766] env[61907]: DEBUG nova.compute.manager [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 891.750124] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.751268] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda34d51-2bd5-4eb4-87d2-b624e8046bea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.764536] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.764953] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f41f32d-27f6-4d7b-b489-15dc772e4b81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.775078] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 891.775078] env[61907]: value = "task-1244028" [ 891.775078] env[61907]: _type = "Task" [ 891.775078] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.784827] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244028, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.813549] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12ac96bb-d837-4c12-8206-f20ffc00e549 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.822599] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897d0367-cf97-42eb-8949-f9377cbe41cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.861069] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb95ed58-c412-407e-93bb-5ffe2a5ff214 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.870664] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a90fc22-73b0-4e0f-ac10-ef12f4c44247 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.888534] env[61907]: DEBUG nova.compute.provider_tree [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 891.974063] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a03852be-1358-4fc4-ac3e-b8f875900534 tempest-ServerRescueTestJSON-860546575 tempest-ServerRescueTestJSON-860546575-project-member] Lock "865d386d-c5f0-439b-b50b-b613993ec9b8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.298s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.008831] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244027, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.055685] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d5ac6a-45de-4f93-f70d-15aef25f11d8, 'name': SearchDatastore_Task, 'duration_secs': 0.011555} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.056200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.056688] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/53176c9e-d15c-49d5-b4a9-22b780279ecb.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.057288] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff66402b-cf95-4c29-b6e2-6acd67b38db0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.068037] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 892.068037] env[61907]: value = "task-1244029" [ 892.068037] env[61907]: _type = "Task" [ 892.068037] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.079308] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244029, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.089655] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.009s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.110563] env[61907]: INFO nova.compute.manager [-] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Took 2.12 seconds to deallocate network for instance. [ 892.118686] env[61907]: DEBUG nova.network.neutron [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Port 6007e3ea-4d8d-4523-b3da-93d26643424f binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 892.160949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "da490943-511b-4776-8f16-4f51c3b055c4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.160949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.160949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "da490943-511b-4776-8f16-4f51c3b055c4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.160949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.160949] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.162699] env[61907]: INFO nova.compute.manager [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Terminating instance [ 892.278823] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updated VIF entry in instance network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.279266] env[61907]: DEBUG nova.network.neutron [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.300634] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244028, 'name': PowerOffVM_Task, 'duration_secs': 0.294343} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.302158] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.302158] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 892.302158] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1965236-929c-4b7a-b0f0-2c6e43093371 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.391922] env[61907]: DEBUG nova.scheduler.client.report [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 892.396126] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.396499] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.396647] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleting the datastore file [datastore2] d32315e1-68b9-49ee-9393-8b2f9cbaf645 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.397317] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed9721d7-eef1-497d-85f8-69340f0cad10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.407356] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 892.407356] env[61907]: value = "task-1244031" [ 892.407356] env[61907]: _type = "Task" [ 892.407356] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.421246] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244031, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.487805] env[61907]: DEBUG nova.network.neutron [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.509491] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244027, 'name': ReconfigVM_Task, 'duration_secs': 0.563832} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.510648] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 691fff4a-9dbd-47f2-8908-69057a2bb6fe/691fff4a-9dbd-47f2-8908-69057a2bb6fe.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.512053] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-174d3333-689b-45e8-95e5-b0a2b17c3a63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.522029] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 892.522029] env[61907]: value = "task-1244032" [ 892.522029] env[61907]: _type = "Task" [ 892.522029] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.533886] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244032, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.587995] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244029, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.624800] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.671431] env[61907]: DEBUG nova.compute.manager [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 892.671431] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 892.671431] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99acbf7-8ceb-4863-9fc2-a958e29cb957 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.681489] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.686030] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60c4ed74-8c4a-47c4-bdc3-15b5419f0290 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.697203] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 892.697203] env[61907]: value = "task-1244033" [ 892.697203] env[61907]: _type = "Task" [ 892.697203] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.714257] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.783948] env[61907]: DEBUG oslo_concurrency.lockutils [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.784285] env[61907]: DEBUG nova.compute.manager [req-ea8870c6-0ab2-467d-b75b-0af1aa8ffe40 req-55189341-e842-4f4d-97a3-b04c41032236 service nova] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Received event network-vif-deleted-6ef9178a-ce1e-4b38-ba1d-398b51e22951 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 892.833092] env[61907]: DEBUG nova.compute.manager [req-f655d6d2-5b5e-44ba-ad58-ac100dff3492 req-786c6cd4-661f-48c8-8c04-24b3e40c9bc6 service nova] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Received event network-vif-deleted-d1082c95-320a-4bba-b83c-c3e5da0d2384 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 892.850399] env[61907]: DEBUG nova.compute.manager [req-382d0f14-f05d-4f58-ade6-f82d83a09b85 req-855ceffe-3f43-4bd7-87d8-fb2775f8ac55 service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Received event network-vif-deleted-ce4e1e2e-3671-4aba-a760-b96db6f21180 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 892.876323] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Successfully updated port: 5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.898320] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.481s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.901481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.279s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.901754] env[61907]: DEBUG nova.objects.instance [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lazy-loading 'resources' on Instance uuid 5c357582-0b55-4de7-968e-4daa3522dda4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.924593] env[61907]: DEBUG oslo_vmware.api [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244031, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281593} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.924939] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.925165] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.925370] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.925623] env[61907]: INFO nova.compute.manager [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Took 1.18 seconds to destroy the instance on the hypervisor. [ 892.925960] env[61907]: DEBUG oslo.service.loopingcall [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.926196] env[61907]: DEBUG nova.compute.manager [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 892.926459] env[61907]: DEBUG nova.network.neutron [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.932685] env[61907]: INFO nova.scheduler.client.report [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance 37ff6c54-6b79-4a9c-bc16-29d974185d3e [ 892.993784] env[61907]: INFO nova.compute.manager [-] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Took 2.23 seconds to deallocate network for instance. [ 893.034966] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244032, 'name': Rename_Task, 'duration_secs': 0.27501} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.036109] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 893.036612] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b9b9282-34b7-4459-a704-cd7333f44f87 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.046918] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 893.046918] env[61907]: value = "task-1244034" [ 893.046918] env[61907]: _type = "Task" [ 893.046918] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.061443] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244034, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.080865] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244029, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544644} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.081249] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/53176c9e-d15c-49d5-b4a9-22b780279ecb.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.082086] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.082086] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8587deb0-83c9-467e-8591-b99c9c9ac445 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.093668] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 893.093668] env[61907]: value = "task-1244035" [ 893.093668] env[61907]: _type = "Task" [ 893.093668] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.106678] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244035, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.152552] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.152850] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.153056] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.193527] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.193857] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.194137] env[61907]: INFO nova.compute.manager [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Attaching volume c4e7f5a7-d519-4e29-b4c1-21dcb964d875 to /dev/sdb [ 893.213844] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244033, 'name': PowerOffVM_Task, 'duration_secs': 0.306378} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.214378] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 893.214378] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 893.214729] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b15e978a-bcd4-4085-b680-06ca83b17b39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.247689] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6672ed0f-df9d-45cf-a9a4-f1ad3ff0f256 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.260334] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10c63cad-9c67-455f-b8cb-f15d14900659 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.283843] env[61907]: DEBUG nova.virt.block_device [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating existing volume attachment record: 9970450f-5d4a-4c36-ae22-3ce1e5a403d9 {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 893.319764] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 893.319764] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 893.319764] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Deleting the datastore file [datastore1] da490943-511b-4776-8f16-4f51c3b055c4 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.320141] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f03096d-d94a-4350-afe2-36098f805097 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.331127] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for the task: (returnval){ [ 893.331127] env[61907]: value = "task-1244037" [ 893.331127] env[61907]: _type = "Task" [ 893.331127] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.346495] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244037, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.382065] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.382201] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquired lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.382456] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.447737] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b65435ab-8fe5-45dc-b96d-2d5a0a26d9d3 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "37ff6c54-6b79-4a9c-bc16-29d974185d3e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.433s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.502553] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.557945] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244034, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.613150] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244035, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103563} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.613150] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.613150] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d5bed4d-f9ff-4d9a-8b7c-62658bc93fcb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.640511] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/53176c9e-d15c-49d5-b4a9-22b780279ecb.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.644274] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afb1c055-ed51-4bb2-be44-9c0184713768 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.668968] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 893.668968] env[61907]: value = "task-1244040" [ 893.668968] env[61907]: _type = "Task" [ 893.668968] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.680750] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.806594] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-490fed12-6235-4cc3-bb8b-67e8707d74dd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.815548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67622548-4063-4213-9104-1958a5d16c8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.861821] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdc6326a-7321-4287-b14f-b937f0ab33b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.873203] env[61907]: DEBUG oslo_vmware.api [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Task: {'id': task-1244037, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17163} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.875888] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.876304] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.876564] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.876821] env[61907]: INFO nova.compute.manager [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Took 1.21 seconds to destroy the instance on the hypervisor. [ 893.877156] env[61907]: DEBUG oslo.service.loopingcall [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.877573] env[61907]: DEBUG nova.compute.manager [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 893.877699] env[61907]: DEBUG nova.network.neutron [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 893.880803] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48462ce-6884-42fc-a990-ba9d07998e30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.903477] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.954278] env[61907]: DEBUG nova.network.neutron [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.980248] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.059039] env[61907]: DEBUG oslo_vmware.api [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244034, 'name': PowerOnVM_Task, 'duration_secs': 0.723723} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.059451] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 894.183813] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244040, 'name': ReconfigVM_Task, 'duration_secs': 0.469733} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.184092] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/53176c9e-d15c-49d5-b4a9-22b780279ecb.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.185055] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-715e003d-a2c2-4a65-9e85-6352366054a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.197578] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 894.197578] env[61907]: value = "task-1244042" [ 894.197578] env[61907]: _type = "Task" [ 894.197578] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.209667] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244042, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.210815] env[61907]: DEBUG nova.compute.manager [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 894.211771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f75538a-569f-490f-91dc-833887fd77ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.292062] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.292062] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.292610] env[61907]: DEBUG nova.network.neutron [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 894.319730] env[61907]: DEBUG nova.network.neutron [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updating instance_info_cache with network_info: [{"id": "5d7e9191-f828-48c6-81f9-0f30c6620477", "address": "fa:16:3e:7e:24:34", "network": {"id": "b7dfc687-b9ee-4774-986a-ef0bca793b1b", "bridge": "br-int", "label": "tempest-ServersTestJSON-762133036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0b2c5e8cc184e328a19e0c0a8cf8169", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d7e9191-f8", "ovs_interfaceid": "5d7e9191-f828-48c6-81f9-0f30c6620477", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.436673] env[61907]: ERROR nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [req-e98f2794-365b-4356-bd27-055bfeb20f67] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e98f2794-365b-4356-bd27-055bfeb20f67"}]} [ 894.459797] env[61907]: INFO nova.compute.manager [-] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Took 1.53 seconds to deallocate network for instance. [ 894.471200] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 894.499441] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 894.499441] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 894.513873] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 894.542956] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 894.727770] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244042, 'name': Rename_Task, 'duration_secs': 0.221925} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.734656] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.734906] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7a6dcaea-12d7-43d2-974e-5b405440c501 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.741000] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c34c9d49-e1e4-4084-8011-c7e1cb1c5304 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 34.736s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.747050] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 894.747050] env[61907]: value = "task-1244043" [ 894.747050] env[61907]: _type = "Task" [ 894.747050] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.756481] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244043, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.823521] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Releasing lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.823857] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Instance network_info: |[{"id": "5d7e9191-f828-48c6-81f9-0f30c6620477", "address": "fa:16:3e:7e:24:34", "network": {"id": "b7dfc687-b9ee-4774-986a-ef0bca793b1b", "bridge": "br-int", "label": "tempest-ServersTestJSON-762133036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0b2c5e8cc184e328a19e0c0a8cf8169", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d7e9191-f8", "ovs_interfaceid": "5d7e9191-f828-48c6-81f9-0f30c6620477", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 894.825093] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:24:34', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '650f7968-4522-4ba5-8304-1b9949951ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5d7e9191-f828-48c6-81f9-0f30c6620477', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.833189] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Creating folder: Project (b0b2c5e8cc184e328a19e0c0a8cf8169). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.839587] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f280138c-a3df-447a-95f6-5a5d796183c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.854842] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Created folder: Project (b0b2c5e8cc184e328a19e0c0a8cf8169) in parent group-v268168. [ 894.855114] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Creating folder: Instances. Parent ref: group-v268279. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 894.858562] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6e87c3d5-6538-46dc-a59f-a8764e58d925 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.860984] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8672e27d-9934-49a6-987c-760aa838e1ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.870529] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36e61f2-022c-4112-a4d7-52b7150e1879 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.875437] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Created folder: Instances in parent group-v268279. [ 894.875710] env[61907]: DEBUG oslo.service.loopingcall [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.876571] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.877218] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-61119750-3d81-4fb5-a6bf-3cc6f2b582ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.920739] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd1c88c-270c-44bb-bbea-e0d30ee82edb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.926454] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.926454] env[61907]: value = "task-1244046" [ 894.926454] env[61907]: _type = "Task" [ 894.926454] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.932722] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844cdb4f-4606-4da1-a140-798ede5cf61b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.947709] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244046, 'name': CreateVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.954082] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 894.975958] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.042506] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Received event network-vif-plugged-5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 895.042506] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Acquiring lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.046277] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.046277] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.046277] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] No waiting events found dispatching network-vif-plugged-5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 895.046277] env[61907]: WARNING nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Received unexpected event network-vif-plugged-5d7e9191-f828-48c6-81f9-0f30c6620477 for instance with vm_state building and task_state spawning. [ 895.046277] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Received event network-changed-5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 895.046277] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Refreshing instance network info cache due to event network-changed-5d7e9191-f828-48c6-81f9-0f30c6620477. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 895.046277] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Acquiring lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.046277] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Acquired lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.046277] env[61907]: DEBUG nova.network.neutron [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Refreshing network info cache for port 5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.082365] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.082631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.131776] env[61907]: DEBUG nova.network.neutron [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.256079] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244043, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.264717] env[61907]: DEBUG nova.network.neutron [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.440100] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244046, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.477942] env[61907]: ERROR nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [req-b6bf23a9-c8ac-42f3-833d-d355c8accf5d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-b6bf23a9-c8ac-42f3-833d-d355c8accf5d"}]} [ 895.497168] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 895.513332] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 895.513627] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 895.527901] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 895.546246] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 895.584783] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 895.635481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.732844] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.732933] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.733146] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.733365] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.733556] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.736060] env[61907]: INFO nova.compute.manager [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Terminating instance [ 895.763608] env[61907]: DEBUG oslo_vmware.api [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244043, 'name': PowerOnVM_Task, 'duration_secs': 0.649931} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.766523] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.766688] env[61907]: INFO nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Took 9.10 seconds to spawn the instance on the hypervisor. [ 895.766882] env[61907]: DEBUG nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 895.767560] env[61907]: INFO nova.compute.manager [-] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Took 1.89 seconds to deallocate network for instance. [ 895.768507] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b708e8be-8a5c-4945-8b53-53ed79f1846f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.839734] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3091ae63-1cbd-4d2b-8ff9-6d687c83dda9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.848556] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b972a94-2b1f-4e96-a5e1-79b0597ed58c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.884027] env[61907]: DEBUG nova.network.neutron [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updated VIF entry in instance network info cache for port 5d7e9191-f828-48c6-81f9-0f30c6620477. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.884426] env[61907]: DEBUG nova.network.neutron [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updating instance_info_cache with network_info: [{"id": "5d7e9191-f828-48c6-81f9-0f30c6620477", "address": "fa:16:3e:7e:24:34", "network": {"id": "b7dfc687-b9ee-4774-986a-ef0bca793b1b", "bridge": "br-int", "label": "tempest-ServersTestJSON-762133036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0b2c5e8cc184e328a19e0c0a8cf8169", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d7e9191-f8", "ovs_interfaceid": "5d7e9191-f828-48c6-81f9-0f30c6620477", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.886138] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab31f0a-460b-49ed-9513-3248c00f471c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.897420] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d07d92-51eb-4954-86fe-eee0c0d4f478 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.912885] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 895.939797] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244046, 'name': CreateVM_Task, 'duration_secs': 0.529582} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.939975] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.940715] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.941324] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.941407] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.941937] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-395d2169-86f2-4619-8f93-857e94eda704 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.948185] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 895.948185] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e3508a-6571-81ad-7b14-3ec4559d60a6" [ 895.948185] env[61907]: _type = "Task" [ 895.948185] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.957072] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e3508a-6571-81ad-7b14-3ec4559d60a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.105926] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.158193] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e15a593e-f112-46df-b946-eb7f05a5fadd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.178773] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82380ea8-bdf4-43a9-85c5-e2bdddaded18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.187176] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 896.240218] env[61907]: DEBUG nova.compute.manager [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 896.240470] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 896.241535] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b003161-37e0-48a9-b333-ba34cd80b45a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.252681] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 896.253022] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67a1dbc5-94fb-401a-888c-b21825280bc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.261955] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 896.261955] env[61907]: value = "task-1244048" [ 896.261955] env[61907]: _type = "Task" [ 896.261955] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.275449] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244048, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.284709] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.296282] env[61907]: INFO nova.compute.manager [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Took 17.06 seconds to build instance. [ 896.393055] env[61907]: DEBUG oslo_concurrency.lockutils [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] Releasing lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.393055] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Received event network-vif-deleted-3da06fb1-6740-4ca1-9765-667fcac56116 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 896.393669] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Received event network-vif-deleted-e34b5242-a01d-49e3-9aba-aaf54bf279df {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 896.393669] env[61907]: INFO nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Neutron deleted interface e34b5242-a01d-49e3-9aba-aaf54bf279df; detaching it from the instance and deleting it from the info cache [ 896.393669] env[61907]: DEBUG nova.network.neutron [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.436723] env[61907]: ERROR nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] [req-c4f8d56c-853d-4ad4-9259-71650aa9e32e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-c4f8d56c-853d-4ad4-9259-71650aa9e32e"}]} [ 896.455652] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 896.466683] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e3508a-6571-81ad-7b14-3ec4559d60a6, 'name': SearchDatastore_Task, 'duration_secs': 0.011635} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.466683] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.467213] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.467586] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.467862] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.468134] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.468441] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-38dc4921-1521-4776-bce7-dc07158a4e9f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.471767] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 896.472102] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 896.483829] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.485290] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.486062] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 896.488291] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38a1af28-6e80-4839-bb99-08b1f7515478 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.496552] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 896.496552] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d40c71-5f5e-18eb-1b8a-a94306e98721" [ 896.496552] env[61907]: _type = "Task" [ 896.496552] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.505186] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d40c71-5f5e-18eb-1b8a-a94306e98721, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.510795] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 896.675540] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "c3bf9001-7166-4064-89d5-96be073fa7c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.676576] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.695541] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.696589] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb29e215-fa24-41e4-b849-d1f1e1ac5d32 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.705016] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 896.705016] env[61907]: value = "task-1244049" [ 896.705016] env[61907]: _type = "Task" [ 896.705016] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.717545] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244049, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.772375] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244048, 'name': PowerOffVM_Task, 'duration_secs': 0.263336} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.772685] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 896.772862] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 896.775582] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-45da29cf-f359-446e-a2e4-b34639224c6f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.789266] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b72e9e7c-563e-4c14-9121-19c16868678a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.798414] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c98c6a-c5f9-46f9-baf7-1f60f3389639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.802376] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5d1fc40a-35a2-4c0c-84bb-280aa7c1f361 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.575s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.835994] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcb1647-a71a-42ac-ad3f-4f1d017ba02d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.845820] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc122213-c548-48ac-b52d-e37289334864 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.852301] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 896.853076] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 896.853076] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleting the datastore file [datastore1] 2c1212a7-df13-49b7-b3b3-24fb59b6789d {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 896.853437] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd1e4df2-7311-4112-b5d0-826d9199153c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.864874] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 896.871037] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 896.871037] env[61907]: value = "task-1244051" [ 896.871037] env[61907]: _type = "Task" [ 896.871037] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.871517] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "71804bc1-79ce-4731-9e27-74c007d4e906" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.871858] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.884241] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244051, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.897868] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ca13c58-69b7-4f5a-bb85-a7c306a5976f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.908543] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44fafdf5-07b4-4625-8fb8-b915bf2b5e56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.942654] env[61907]: DEBUG nova.compute.manager [req-0da8f4f2-80cf-4d24-90bd-c70efd099e04 req-58b637e8-49f2-4002-bf2f-8223fc64a36a service nova] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Detach interface failed, port_id=e34b5242-a01d-49e3-9aba-aaf54bf279df, reason: Instance da490943-511b-4776-8f16-4f51c3b055c4 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 897.008170] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d40c71-5f5e-18eb-1b8a-a94306e98721, 'name': SearchDatastore_Task, 'duration_secs': 0.011611} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.009173] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2cd533f-66c6-4309-b3eb-c553a9e420d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.015428] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 897.015428] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5268ab3e-9f64-3aa7-b9e8-cc014f2b21fc" [ 897.015428] env[61907]: _type = "Task" [ 897.015428] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.025677] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5268ab3e-9f64-3aa7-b9e8-cc014f2b21fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.182071] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 897.216965] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244049, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.354637] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 897.355463] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132b0ae5-12aa-4185-83ff-5091ca19ef83 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.363366] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 897.363597] env[61907]: ERROR oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk due to incomplete transfer. [ 897.363865] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-7fe9ddd4-c963-4fc4-8261-4b7d876588f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.380400] env[61907]: DEBUG oslo_vmware.rw_handles [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522c5391-4284-2009-9e0f-1fd89024fbe1/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 897.380640] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Uploaded image 93afda2d-a435-4b9d-983e-30d2ad798313 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 897.382861] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 897.383238] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 897.385857] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8ce8e869-9b94-4227-8bc1-63470114ed69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.395895] env[61907]: DEBUG oslo_vmware.api [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244051, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153806} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.397527] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.397747] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 897.397935] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 897.398237] env[61907]: INFO nova.compute.manager [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 897.398413] env[61907]: DEBUG oslo.service.loopingcall [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.398714] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 897.398714] env[61907]: value = "task-1244052" [ 897.398714] env[61907]: _type = "Task" [ 897.398714] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.398912] env[61907]: DEBUG nova.compute.manager [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 897.399015] env[61907]: DEBUG nova.network.neutron [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 897.406604] env[61907]: DEBUG nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 897.406861] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 95 to 96 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 897.407061] env[61907]: DEBUG nova.compute.provider_tree [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.418807] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244052, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.526813] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5268ab3e-9f64-3aa7-b9e8-cc014f2b21fc, 'name': SearchDatastore_Task, 'duration_secs': 0.010807} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.527210] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.527278] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] a31110ed-c4bb-4492-95a7-0a0ef5010c28/a31110ed-c4bb-4492-95a7-0a0ef5010c28.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.527532] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23e30aff-1d36-48d4-b65a-fd0ebeefeb63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.535139] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 897.535139] env[61907]: value = "task-1244053" [ 897.535139] env[61907]: _type = "Task" [ 897.535139] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.547790] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244053, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.712584] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.720931] env[61907]: DEBUG oslo_vmware.api [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244049, 'name': PowerOnVM_Task, 'duration_secs': 0.812721} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.721286] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.721522] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-77a2fd84-f7a4-40df-af0c-f17ca9a7744c tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance '0bfdb2d0-8388-4be8-a2ee-743c029db6c3' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 897.786424] env[61907]: DEBUG nova.compute.manager [req-36e1b02c-f4ec-47fe-875f-777af8481f1f req-e1ee6985-3b40-46fe-8ab5-75d6f7060588 service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Received event network-vif-deleted-e0833683-f8a7-41f3-bca6-307d94fffbfd {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 897.786642] env[61907]: INFO nova.compute.manager [req-36e1b02c-f4ec-47fe-875f-777af8481f1f req-e1ee6985-3b40-46fe-8ab5-75d6f7060588 service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Neutron deleted interface e0833683-f8a7-41f3-bca6-307d94fffbfd; detaching it from the instance and deleting it from the info cache [ 897.786853] env[61907]: DEBUG nova.network.neutron [req-36e1b02c-f4ec-47fe-875f-777af8481f1f req-e1ee6985-3b40-46fe-8ab5-75d6f7060588 service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.885370] env[61907]: DEBUG nova.compute.manager [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 897.885655] env[61907]: DEBUG nova.compute.manager [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing instance network info cache due to event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 897.886081] env[61907]: DEBUG oslo_concurrency.lockutils [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.886291] env[61907]: DEBUG oslo_concurrency.lockutils [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.886608] env[61907]: DEBUG nova.network.neutron [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.888938] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Volume attach. Driver type: vmdk {{(pid=61907) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 897.889211] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268278', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'name': 'volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4c70ef23-fd26-4e90-b227-13aa050ff46d', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'serial': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 897.894128] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c424be10-df78-4654-adb6-454a921a13b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.919351] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 5.018s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.922588] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.926512] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.425s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.926796] env[61907]: DEBUG nova.objects.instance [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'resources' on Instance uuid 52404147-0375-4f9c-9e1e-4a52001df7cb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.928802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a58a0d58-89db-4e8e-918a-3a393d51d3c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.970533] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875/volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.974832] env[61907]: INFO nova.scheduler.client.report [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Deleted allocations for instance 5c357582-0b55-4de7-968e-4daa3522dda4 [ 897.976267] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-37b185d9-4e06-45ff-b918-85c9d334e755 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.990117] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244052, 'name': Destroy_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.003512] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 898.003512] env[61907]: value = "task-1244054" [ 898.003512] env[61907]: _type = "Task" [ 898.003512] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.015312] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244054, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.047299] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244053, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.257462] env[61907]: DEBUG nova.network.neutron [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.298132] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef95f763-8ef3-4744-ba6e-cec17c9559fe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.309478] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72a2c7bc-99de-44ac-bfa5-57c44e4439d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.344822] env[61907]: DEBUG nova.compute.manager [req-36e1b02c-f4ec-47fe-875f-777af8481f1f req-e1ee6985-3b40-46fe-8ab5-75d6f7060588 service nova] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Detach interface failed, port_id=e0833683-f8a7-41f3-bca6-307d94fffbfd, reason: Instance 2c1212a7-df13-49b7-b3b3-24fb59b6789d could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 898.432915] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244052, 'name': Destroy_Task, 'duration_secs': 0.66693} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.436795] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Destroyed the VM [ 898.440579] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 898.440579] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e17e96f7-9909-4bdd-bbb1-685c573292ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.448311] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 898.448311] env[61907]: value = "task-1244055" [ 898.448311] env[61907]: _type = "Task" [ 898.448311] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.466935] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.500071] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7df1aca2-32ef-4337-aaf5-8910e9574771 tempest-ServerGroupTestJSON-1506054560 tempest-ServerGroupTestJSON-1506054560-project-member] Lock "5c357582-0b55-4de7-968e-4daa3522dda4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.163s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.521110] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244054, 'name': ReconfigVM_Task, 'duration_secs': 0.492692} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.521110] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875/volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.526823] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d2d5d07b-b019-497b-81eb-247b7c1a0928 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.548746] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 898.548746] env[61907]: value = "task-1244056" [ 898.548746] env[61907]: _type = "Task" [ 898.548746] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.552227] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244053, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53267} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.555457] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] a31110ed-c4bb-4492-95a7-0a0ef5010c28/a31110ed-c4bb-4492-95a7-0a0ef5010c28.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.555738] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.558377] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04afd20c-a108-40d6-b520-6a2560c7acd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.568592] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244056, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.570296] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 898.570296] env[61907]: value = "task-1244057" [ 898.570296] env[61907]: _type = "Task" [ 898.570296] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.584399] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.732442] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e43ee5-3855-4e84-8df7-9ac24477f693 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.748044] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51a5c4f4-ec77-49a4-8857-f5751957dab5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.784913] env[61907]: INFO nova.compute.manager [-] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Took 1.39 seconds to deallocate network for instance. [ 898.790248] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9ebe0ec-af0d-4d56-a330-dd7949303e53 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.804016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a1dae7f-9ffc-423a-a531-e3e7733a3d30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.820181] env[61907]: DEBUG nova.compute.provider_tree [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.946644] env[61907]: DEBUG nova.network.neutron [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updated VIF entry in instance network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.947039] env[61907]: DEBUG nova.network.neutron [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.961932] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.061632] env[61907]: DEBUG oslo_vmware.api [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244056, 'name': ReconfigVM_Task, 'duration_secs': 0.235771} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.061960] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268278', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'name': 'volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4c70ef23-fd26-4e90-b227-13aa050ff46d', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'serial': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 899.080358] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186916} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.080635] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 899.081416] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52fdc0c-6220-4bb3-8a27-d30feaa8c6ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.104013] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] a31110ed-c4bb-4492-95a7-0a0ef5010c28/a31110ed-c4bb-4492-95a7-0a0ef5010c28.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 899.104303] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6d56ef6-30d6-44e7-ab23-72c3614f3ba9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.127249] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 899.127249] env[61907]: value = "task-1244058" [ 899.127249] env[61907]: _type = "Task" [ 899.127249] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.135128] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244058, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.300902] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.323158] env[61907]: DEBUG nova.scheduler.client.report [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 899.440974] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "23f9f3d1-b376-4502-927c-e9bd89693131" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.441273] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.449327] env[61907]: DEBUG oslo_concurrency.lockutils [req-569461e2-9ea0-4852-9a68-eeffca4596bd req-c265b3bb-fd7b-4967-b7b2-ae357680d495 service nova] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.463711] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244055, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.638090] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244058, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.829707] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.832896] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.857s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.833392] env[61907]: DEBUG nova.objects.instance [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lazy-loading 'resources' on Instance uuid d32315e1-68b9-49ee-9393-8b2f9cbaf645 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.860345] env[61907]: INFO nova.scheduler.client.report [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted allocations for instance 52404147-0375-4f9c-9e1e-4a52001df7cb [ 899.944780] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 899.968021] env[61907]: DEBUG oslo_vmware.api [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244055, 'name': RemoveSnapshot_Task, 'duration_secs': 1.435867} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.968021] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 899.968021] env[61907]: INFO nova.compute.manager [None req-4ef3a697-fe95-4bab-9490-2e825a375b92 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Took 16.61 seconds to snapshot the instance on the hypervisor. [ 900.107129] env[61907]: DEBUG nova.objects.instance [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.141283] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244058, 'name': ReconfigVM_Task, 'duration_secs': 0.62179} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.141578] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Reconfigured VM instance instance-0000004f to attach disk [datastore2] a31110ed-c4bb-4492-95a7-0a0ef5010c28/a31110ed-c4bb-4492-95a7-0a0ef5010c28.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 900.143618] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9fbc854b-89d4-489b-a4ac-5669e2365848 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.152020] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 900.152020] env[61907]: value = "task-1244059" [ 900.152020] env[61907]: _type = "Task" [ 900.152020] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.164281] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244059, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.371012] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3dc105d3-229e-4939-b99e-fdca62ccf5cb tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "52404147-0375-4f9c-9e1e-4a52001df7cb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.067s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.373270] env[61907]: DEBUG oslo_concurrency.lockutils [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] Acquired lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.374406] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0561b1be-b5ab-4630-8ebb-7f4ae49d6584 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.384915] env[61907]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 900.385089] env[61907]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61907) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 900.385533] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ee950d4-75d6-4cb6-8eaa-ce8ff8de9cf3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.402410] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29999779-c8c5-4811-9874-4a513c672171 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.448956] env[61907]: ERROR root [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-268220' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-268220' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-268220' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-268220'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-268220' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-268220' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-268220'}\n"]: nova.exception.InstanceNotFound: Instance 52404147-0375-4f9c-9e1e-4a52001df7cb could not be found. [ 900.449250] env[61907]: DEBUG oslo_concurrency.lockutils [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] Releasing lock "52404147-0375-4f9c-9e1e-4a52001df7cb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 900.449467] env[61907]: DEBUG nova.compute.manager [req-308ee2c4-a58c-4827-a085-46e66ac74b47 req-c3aa5796-b22b-45ed-ab2e-1afe699b874a service nova] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Detach interface failed, port_id=b0ff98c3-b17f-497e-8ba2-0578ebbd2927, reason: Instance 52404147-0375-4f9c-9e1e-4a52001df7cb could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 900.472874] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.524713] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.524987] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.525199] env[61907]: DEBUG nova.compute.manager [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Going to confirm migration 1 {{(pid=61907) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 900.614928] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b57348a-89ad-42f1-95e7-3d54008c216b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.421s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.663301] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244059, 'name': Rename_Task, 'duration_secs': 0.294193} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.666672] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.668491] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef0fc41e-b7b9-4bd9-8768-c9ca8ff431a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.677079] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 900.677079] env[61907]: value = "task-1244060" [ 900.677079] env[61907]: _type = "Task" [ 900.677079] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.690098] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244060, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.699328] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52fce6c-b6fb-4901-9130-d6ea544f1927 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.710146] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a589c6f1-3f9e-47f0-9f88-f0ab21092ebb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.742537] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c66e72a-907a-4648-97ba-faa5cd70ca35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.751965] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb802d9-481d-4b5d-b6f3-d61c883a1b37 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.756509] env[61907]: DEBUG oslo_concurrency.lockutils [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.756667] env[61907]: DEBUG oslo_concurrency.lockutils [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.756855] env[61907]: DEBUG nova.compute.manager [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 900.757669] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e62c4ab1-8939-4661-8c6b-475dfa7065bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.775351] env[61907]: DEBUG nova.compute.provider_tree [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.778752] env[61907]: DEBUG nova.compute.manager [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 900.782052] env[61907]: DEBUG nova.objects.instance [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.096507] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.096852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.097175] env[61907]: DEBUG nova.network.neutron [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 901.097497] env[61907]: DEBUG nova.objects.instance [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'info_cache' on Instance uuid 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.188464] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244060, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.285078] env[61907]: DEBUG nova.scheduler.client.report [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 901.688285] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244060, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.793016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.795435] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.796069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.690s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.797554] env[61907]: INFO nova.compute.claims [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 901.800110] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1826902-db06-49d0-9911-4ad2754eb025 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.810051] env[61907]: DEBUG oslo_vmware.api [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 901.810051] env[61907]: value = "task-1244061" [ 901.810051] env[61907]: _type = "Task" [ 901.810051] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.819607] env[61907]: DEBUG oslo_vmware.api [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244061, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.820756] env[61907]: INFO nova.scheduler.client.report [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted allocations for instance d32315e1-68b9-49ee-9393-8b2f9cbaf645 [ 902.188339] env[61907]: DEBUG oslo_vmware.api [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244060, 'name': PowerOnVM_Task, 'duration_secs': 1.045661} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.190599] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.190799] env[61907]: INFO nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Took 10.85 seconds to spawn the instance on the hypervisor. [ 902.191107] env[61907]: DEBUG nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.191965] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5a928f-e0a3-4db3-bdf3-be46edabcce1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.311210] env[61907]: DEBUG nova.network.neutron [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [{"id": "6007e3ea-4d8d-4523-b3da-93d26643424f", "address": "fa:16:3e:01:4f:1c", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6007e3ea-4d", "ovs_interfaceid": "6007e3ea-4d8d-4523-b3da-93d26643424f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.323225] env[61907]: DEBUG oslo_vmware.api [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244061, 'name': PowerOffVM_Task, 'duration_secs': 0.264066} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.323465] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.323679] env[61907]: DEBUG nova.compute.manager [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 902.324674] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-493f104a-557a-417b-9728-f18683349094 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.333909] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c1c34415-3edd-41ef-be98-017ca75e7851 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "d32315e1-68b9-49ee-9393-8b2f9cbaf645" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.092s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.499353] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.499353] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.711159] env[61907]: INFO nova.compute.manager [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Took 20.06 seconds to build instance. [ 902.816685] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-0bfdb2d0-8388-4be8-a2ee-743c029db6c3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.817020] env[61907]: DEBUG nova.objects.instance [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'migration_context' on Instance uuid 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.838273] env[61907]: DEBUG oslo_concurrency.lockutils [None req-241c2f12-9f6d-474a-9278-266a5396865f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.081s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.001640] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 903.087951] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64395303-167f-4e59-8462-8a6786ad1aa9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.097418] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff93fa1-ec95-496b-a9f3-13cf7fd85f0c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.132364] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdba9b0-b0d5-4cee-89e6-3c3b762de21f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.143749] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4f7bf70-6e5f-4087-8f95-380c56be1764 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.166602] env[61907]: DEBUG nova.compute.provider_tree [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.184347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "cd2c70a7-a290-4d1a-91fe-5da772860603" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.184778] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.213575] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2701200e-4869-4a17-bace-17d04bdb18ac tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.566s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.320082] env[61907]: DEBUG nova.objects.base [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Object Instance<0bfdb2d0-8388-4be8-a2ee-743c029db6c3> lazy-loaded attributes: info_cache,migration_context {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 903.322260] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14546cf9-97f1-48e7-bd65-68341ab70568 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.342037] env[61907]: DEBUG nova.objects.instance [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.344419] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12c35cbb-c9e6-48ec-8e75-71256119c6a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.352124] env[61907]: DEBUG oslo_vmware.api [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 903.352124] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5249426c-5e2a-fe89-af8e-a87038bb5be7" [ 903.352124] env[61907]: _type = "Task" [ 903.352124] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.359407] env[61907]: DEBUG oslo_vmware.api [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5249426c-5e2a-fe89-af8e-a87038bb5be7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.527523] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.618292] env[61907]: DEBUG nova.compute.manager [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Received event network-changed-5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 903.618502] env[61907]: DEBUG nova.compute.manager [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Refreshing instance network info cache due to event network-changed-5d7e9191-f828-48c6-81f9-0f30c6620477. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 903.619341] env[61907]: DEBUG oslo_concurrency.lockutils [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] Acquiring lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.619477] env[61907]: DEBUG oslo_concurrency.lockutils [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] Acquired lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.619667] env[61907]: DEBUG nova.network.neutron [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Refreshing network info cache for port 5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 903.669691] env[61907]: DEBUG nova.scheduler.client.report [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 903.687701] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 903.847513] env[61907]: DEBUG oslo_concurrency.lockutils [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.847776] env[61907]: DEBUG oslo_concurrency.lockutils [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.847886] env[61907]: DEBUG nova.network.neutron [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 903.848096] env[61907]: DEBUG nova.objects.instance [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'info_cache' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 903.862899] env[61907]: DEBUG oslo_vmware.api [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5249426c-5e2a-fe89-af8e-a87038bb5be7, 'name': SearchDatastore_Task, 'duration_secs': 0.009315} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.863974] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.176304] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.380s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.176988] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 904.180207] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.895s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.180207] env[61907]: DEBUG nova.objects.instance [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lazy-loading 'resources' on Instance uuid da490943-511b-4776-8f16-4f51c3b055c4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.210868] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.355476] env[61907]: DEBUG nova.objects.base [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Object Instance<4c70ef23-fd26-4e90-b227-13aa050ff46d> lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 904.375958] env[61907]: DEBUG nova.network.neutron [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updated VIF entry in instance network info cache for port 5d7e9191-f828-48c6-81f9-0f30c6620477. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 904.376345] env[61907]: DEBUG nova.network.neutron [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updating instance_info_cache with network_info: [{"id": "5d7e9191-f828-48c6-81f9-0f30c6620477", "address": "fa:16:3e:7e:24:34", "network": {"id": "b7dfc687-b9ee-4774-986a-ef0bca793b1b", "bridge": "br-int", "label": "tempest-ServersTestJSON-762133036-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b0b2c5e8cc184e328a19e0c0a8cf8169", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5d7e9191-f8", "ovs_interfaceid": "5d7e9191-f828-48c6-81f9-0f30c6620477", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.683376] env[61907]: DEBUG nova.compute.utils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 904.688602] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 904.688602] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 904.740093] env[61907]: DEBUG nova.policy [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 904.879659] env[61907]: DEBUG oslo_concurrency.lockutils [req-fb79067e-33c6-4322-982d-17d12aba5ceb req-40f482e3-83ab-4fa0-94b6-0bd2e533908e service nova] Releasing lock "refresh_cache-a31110ed-c4bb-4492-95a7-0a0ef5010c28" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.951335] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cd1ad1-9e33-4139-badb-6c4b5967c09a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.961768] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292b7ed2-ad2a-4bab-9d05-1ab32b8eef2b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.998390] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30471861-0305-44f9-a781-c935c781082a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.007089] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc9670f6-7e40-4c3d-a04b-a4c4cfa20fbf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.024693] env[61907]: DEBUG nova.compute.provider_tree [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.110439] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Successfully created port: 61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 905.188796] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 905.274892] env[61907]: DEBUG nova.network.neutron [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.529473] env[61907]: DEBUG nova.scheduler.client.report [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 905.778384] env[61907]: DEBUG oslo_concurrency.lockutils [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.034283] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.854s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.036703] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.324s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.038156] env[61907]: INFO nova.compute.claims [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.057075] env[61907]: INFO nova.scheduler.client.report [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Deleted allocations for instance da490943-511b-4776-8f16-4f51c3b055c4 [ 906.198361] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 906.223121] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.223355] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.223540] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.223748] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.223906] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.224078] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.224299] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.224468] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.224672] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.224855] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.225050] env[61907]: DEBUG nova.virt.hardware [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.226023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1de87c-65b6-4b24-941c-79422f1336d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.234518] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4fd56d-998c-4f2c-8a91-0457a5c29ca3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.567066] env[61907]: DEBUG oslo_concurrency.lockutils [None req-89764e96-66f2-4707-b066-49a73b54c237 tempest-AttachInterfacesUnderV243Test-1207551556 tempest-AttachInterfacesUnderV243Test-1207551556-project-member] Lock "da490943-511b-4776-8f16-4f51c3b055c4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.408s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.596580] env[61907]: DEBUG nova.compute.manager [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Received event network-vif-plugged-61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 906.596832] env[61907]: DEBUG oslo_concurrency.lockutils [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] Acquiring lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.597066] env[61907]: DEBUG oslo_concurrency.lockutils [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.597243] env[61907]: DEBUG oslo_concurrency.lockutils [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.597432] env[61907]: DEBUG nova.compute.manager [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] No waiting events found dispatching network-vif-plugged-61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 906.597580] env[61907]: WARNING nova.compute.manager [req-936abb56-9d0d-4d7e-a263-8a10991031f0 req-3b7856cf-0ff3-4993-9ad1-4e6b4dc3a51d service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Received unexpected event network-vif-plugged-61efaa0f-fae2-4091-97b1-7937f954ff49 for instance with vm_state building and task_state spawning. [ 906.686061] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Successfully updated port: 61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 906.784851] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 906.785231] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d13baaa3-a932-4501-aa8b-ee5aead72862 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.793810] env[61907]: DEBUG oslo_vmware.api [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 906.793810] env[61907]: value = "task-1244062" [ 906.793810] env[61907]: _type = "Task" [ 906.793810] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.803501] env[61907]: DEBUG oslo_vmware.api [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244062, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.189755] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.190071] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.190269] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 907.303622] env[61907]: DEBUG oslo_vmware.api [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244062, 'name': PowerOnVM_Task, 'duration_secs': 0.428826} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.308019] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.308019] env[61907]: DEBUG nova.compute.manager [None req-162ebb29-a569-44c8-b600-5083ef98c06e tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 907.308019] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b46beb-b71d-49c5-ac2d-14aaa316d170 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.319542] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a946018-9c24-49c2-a96b-fd39d033e5dd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.328313] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae768396-28ec-468a-8e42-4355d1882668 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.360141] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c09bbd-2bb6-4c50-a120-e2f98358c953 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.369440] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d35b892-cb45-4272-b5c4-a8f10815b3ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.384943] env[61907]: DEBUG nova.compute.provider_tree [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.739291] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 907.888179] env[61907]: DEBUG nova.scheduler.client.report [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 907.946930] env[61907]: DEBUG nova.network.neutron [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Updating instance_info_cache with network_info: [{"id": "61efaa0f-fae2-4091-97b1-7937f954ff49", "address": "fa:16:3e:b0:d0:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61efaa0f-fa", "ovs_interfaceid": "61efaa0f-fae2-4091-97b1-7937f954ff49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.396118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.396385] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 908.398961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.476s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.400240] env[61907]: INFO nova.compute.claims [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.449201] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.449397] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Instance network_info: |[{"id": "61efaa0f-fae2-4091-97b1-7937f954ff49", "address": "fa:16:3e:b0:d0:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61efaa0f-fa", "ovs_interfaceid": "61efaa0f-fae2-4091-97b1-7937f954ff49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 908.449827] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:d0:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61efaa0f-fae2-4091-97b1-7937f954ff49', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.457588] env[61907]: DEBUG oslo.service.loopingcall [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.458654] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 908.458864] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0cb8748-68e6-4339-863b-e2ea4998cb89 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.486028] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.486028] env[61907]: value = "task-1244063" [ 908.486028] env[61907]: _type = "Task" [ 908.486028] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.497929] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244063, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.630066] env[61907]: DEBUG nova.compute.manager [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Received event network-changed-61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 908.630066] env[61907]: DEBUG nova.compute.manager [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Refreshing instance network info cache due to event network-changed-61efaa0f-fae2-4091-97b1-7937f954ff49. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 908.630066] env[61907]: DEBUG oslo_concurrency.lockutils [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] Acquiring lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.630485] env[61907]: DEBUG oslo_concurrency.lockutils [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] Acquired lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.630485] env[61907]: DEBUG nova.network.neutron [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Refreshing network info cache for port 61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 908.905236] env[61907]: DEBUG nova.compute.utils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.906863] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 908.997100] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244063, 'name': CreateVM_Task, 'duration_secs': 0.354226} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.997283] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 908.998014] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.998196] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.998549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.998824] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4accbfae-43d3-4969-b79c-fd752b7b559e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.004692] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 909.004692] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5222f3c0-d607-bedf-131c-4f32d2f49a60" [ 909.004692] env[61907]: _type = "Task" [ 909.004692] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.014082] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5222f3c0-d607-bedf-131c-4f32d2f49a60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.410958] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 909.423096] env[61907]: DEBUG nova.network.neutron [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Updated VIF entry in instance network info cache for port 61efaa0f-fae2-4091-97b1-7937f954ff49. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 909.423490] env[61907]: DEBUG nova.network.neutron [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Updating instance_info_cache with network_info: [{"id": "61efaa0f-fae2-4091-97b1-7937f954ff49", "address": "fa:16:3e:b0:d0:1f", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61efaa0f-fa", "ovs_interfaceid": "61efaa0f-fae2-4091-97b1-7937f954ff49", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.520573] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5222f3c0-d607-bedf-131c-4f32d2f49a60, 'name': SearchDatastore_Task, 'duration_secs': 0.015012} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.521051] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.521345] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.521700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.521880] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.522220] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.525531] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-511aa799-7d05-425a-bf83-563d1294a5d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.538643] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.538846] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 909.539809] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e319a3e-c0fd-455d-a8ee-b0c2a3e92b49 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.549506] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 909.549506] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b2f29-f158-f1b0-7f26-2b92b83a18aa" [ 909.549506] env[61907]: _type = "Task" [ 909.549506] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.560778] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b2f29-f158-f1b0-7f26-2b92b83a18aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.713119] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf0ccec-7391-4ab5-bf11-cc7663feaeee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.721544] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d5d6396-b4d6-4a1f-b8e5-b628f6d441d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.752256] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0780af3e-2a63-4323-8f41-2ebece6ade77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.760395] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f792e1a-0f42-460a-abb2-948cc6cc3585 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.774533] env[61907]: DEBUG nova.compute.provider_tree [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.926532] env[61907]: DEBUG oslo_concurrency.lockutils [req-b8205e0f-ef9e-4aad-8686-bc58dc37b907 req-8781cfac-9e31-457c-93de-676cb0972419 service nova] Releasing lock "refresh_cache-4b07210f-8005-481f-b7a1-ce80a4cbe980" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.064846] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b2f29-f158-f1b0-7f26-2b92b83a18aa, 'name': SearchDatastore_Task, 'duration_secs': 0.012242} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.066495] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0bd79fd-9945-4a11-9c11-e90def2eae12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.073216] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 910.073216] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52df7260-61ae-c379-5ff4-f6c19ba048a4" [ 910.073216] env[61907]: _type = "Task" [ 910.073216] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.083808] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52df7260-61ae-c379-5ff4-f6c19ba048a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.278582] env[61907]: DEBUG nova.scheduler.client.report [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 910.422864] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 910.449682] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.449898] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.450084] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.450283] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.450437] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.450663] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.450922] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.451106] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.451283] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.451477] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.451650] env[61907]: DEBUG nova.virt.hardware [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.452604] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f227d9-8243-4c7b-af62-ce81b90f4a00 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.460743] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb3b816-884d-4c97-8698-773be7985ddc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.475751] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.481182] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Creating folder: Project (48639cd7ada6463299ed2ab4febac821). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.481461] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5b565db-5820-4095-8077-91885b9e00db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.492517] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Created folder: Project (48639cd7ada6463299ed2ab4febac821) in parent group-v268168. [ 910.492713] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Creating folder: Instances. Parent ref: group-v268283. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.492949] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-12f5b4a4-9401-43d6-a1ec-9f6a85611a1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.501656] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Created folder: Instances in parent group-v268283. [ 910.501882] env[61907]: DEBUG oslo.service.loopingcall [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.502099] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.502299] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-254cdc70-b91a-4034-8896-6a953b9b32fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.513817] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 910.514128] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 910.519124] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.519124] env[61907]: value = "task-1244066" [ 910.519124] env[61907]: _type = "Task" [ 910.519124] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.526198] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244066, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.583892] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52df7260-61ae-c379-5ff4-f6c19ba048a4, 'name': SearchDatastore_Task, 'duration_secs': 0.026188} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.584088] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.584386] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 4b07210f-8005-481f-b7a1-ce80a4cbe980/4b07210f-8005-481f-b7a1-ce80a4cbe980.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 910.584664] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbaae730-f247-40d5-ba62-d089ac89d125 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.591044] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 910.591044] env[61907]: value = "task-1244067" [ 910.591044] env[61907]: _type = "Task" [ 910.591044] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.599211] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244067, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.786379] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.387s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.786937] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 910.789631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.489s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.789866] env[61907]: DEBUG nova.objects.instance [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'resources' on Instance uuid 2c1212a7-df13-49b7-b3b3-24fb59b6789d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 911.031433] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244066, 'name': CreateVM_Task, 'duration_secs': 0.374707} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.031652] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.032174] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.032316] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.032680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.032961] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb676142-80aa-4bb1-b3b7-d459b80c62a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.040158] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 911.040158] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d61a2-6ae2-a10a-45ac-5e73162c6115" [ 911.040158] env[61907]: _type = "Task" [ 911.040158] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.048443] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d61a2-6ae2-a10a-45ac-5e73162c6115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.103767] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244067, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.497579} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.104194] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 4b07210f-8005-481f-b7a1-ce80a4cbe980/4b07210f-8005-481f-b7a1-ce80a4cbe980.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 911.104521] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.104882] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5a5634f-d88a-4171-a2c4-dc006aaec77a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.112956] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 911.112956] env[61907]: value = "task-1244068" [ 911.112956] env[61907]: _type = "Task" [ 911.112956] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.123283] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244068, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.293305] env[61907]: DEBUG nova.compute.utils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.295018] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 911.519611] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d54b994-086d-4c23-a1fe-25e6a7061872 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.527483] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7346a76f-aa76-4442-93e7-49bec394c515 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.561366] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee9d4fd6-661e-4d52-b332-c223237cbbf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.568610] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520d61a2-6ae2-a10a-45ac-5e73162c6115, 'name': SearchDatastore_Task, 'duration_secs': 0.032256} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.570700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.570979] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.571255] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.571410] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.571595] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.571886] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e6d951e-1ba3-4318-a7c6-2fa1f3293456 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.574547] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ad3a95-3443-4659-b964-30ffc1b27fce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.588077] env[61907]: DEBUG nova.compute.provider_tree [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.590091] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.590216] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.591149] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eac9e706-4092-4bb3-8dca-c77bfc8a3889 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.596662] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 911.596662] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a209f5-4daf-afb8-99d5-5424df879fcd" [ 911.596662] env[61907]: _type = "Task" [ 911.596662] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.605185] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a209f5-4daf-afb8-99d5-5424df879fcd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.621131] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244068, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10449} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.624016] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.624016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002d53a6-be58-4d70-b750-9e40f9f738f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.643841] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 4b07210f-8005-481f-b7a1-ce80a4cbe980/4b07210f-8005-481f-b7a1-ce80a4cbe980.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.644208] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-93522339-4260-49e1-bb23-344220be5f25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.664802] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 911.664802] env[61907]: value = "task-1244069" [ 911.664802] env[61907]: _type = "Task" [ 911.664802] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.672982] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244069, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.799450] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 912.108346] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a209f5-4daf-afb8-99d5-5424df879fcd, 'name': SearchDatastore_Task, 'duration_secs': 0.009499} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.109126] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-791a951b-51ea-4cc5-8d6a-ea16e4ea5573 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.111830] env[61907]: ERROR nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [req-dd8e7369-bb48-48f7-b6bb-9dcffe09c0f8] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-dd8e7369-bb48-48f7-b6bb-9dcffe09c0f8"}]} [ 912.116900] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 912.116900] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bf8a77-e135-703d-6541-826f9361c67e" [ 912.116900] env[61907]: _type = "Task" [ 912.116900] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.124138] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bf8a77-e135-703d-6541-826f9361c67e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.127218] env[61907]: DEBUG nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 912.141131] env[61907]: DEBUG nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 912.141400] env[61907]: DEBUG nova.compute.provider_tree [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.156182] env[61907]: DEBUG nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 912.174445] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244069, 'name': ReconfigVM_Task, 'duration_secs': 0.312851} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.174445] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 4b07210f-8005-481f-b7a1-ce80a4cbe980/4b07210f-8005-481f-b7a1-ce80a4cbe980.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.175086] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63595bbe-040b-4f08-9026-ab8e188bcbb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.177257] env[61907]: DEBUG nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 912.183936] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 912.183936] env[61907]: value = "task-1244070" [ 912.183936] env[61907]: _type = "Task" [ 912.183936] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.195422] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244070, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.402142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d435ac-8959-41b1-8445-51f119515736 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.410815] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3e8594-c53f-46a5-b68b-b9532f217e82 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.440524] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b82cce8-24d1-43f6-a330-aa50a81654ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.447580] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ad80e0-24a8-423d-9b40-e8f37ea7e94d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.460768] env[61907]: DEBUG nova.compute.provider_tree [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.627305] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bf8a77-e135-703d-6541-826f9361c67e, 'name': SearchDatastore_Task, 'duration_secs': 0.00833} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.627667] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.627832] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] c3bf9001-7166-4064-89d5-96be073fa7c2/c3bf9001-7166-4064-89d5-96be073fa7c2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.628097] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dffba8f5-6b0c-439f-8b78-7f784e218e56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.634429] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 912.634429] env[61907]: value = "task-1244071" [ 912.634429] env[61907]: _type = "Task" [ 912.634429] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.642909] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.693896] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244070, 'name': Rename_Task, 'duration_secs': 0.139659} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.694278] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 912.694548] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba581a50-74e6-4457-aaa8-d9ea32b009c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.700877] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 912.700877] env[61907]: value = "task-1244072" [ 912.700877] env[61907]: _type = "Task" [ 912.700877] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.708497] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.809442] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 912.835364] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.835655] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.835823] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.836032] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.836195] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.836495] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.836749] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.836920] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.837149] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.837393] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.837570] env[61907]: DEBUG nova.virt.hardware [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.838543] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e419c97-1373-4164-a5bf-9d3d6c00c819 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.847630] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627a12e6-10fc-4965-bf5e-54f413070eaa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.862193] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.868159] env[61907]: DEBUG oslo.service.loopingcall [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.868504] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 912.868745] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1f4f5ebd-e36c-45ca-a53f-407d24dae52e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.886659] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.886659] env[61907]: value = "task-1244073" [ 912.886659] env[61907]: _type = "Task" [ 912.886659] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.895315] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244073, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.003289] env[61907]: DEBUG nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 913.003771] env[61907]: DEBUG nova.compute.provider_tree [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 99 to 100 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 913.004049] env[61907]: DEBUG nova.compute.provider_tree [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.145391] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244071, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472401} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.145673] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] c3bf9001-7166-4064-89d5-96be073fa7c2/c3bf9001-7166-4064-89d5-96be073fa7c2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.145902] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.146183] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ffcc397-9828-4047-ab5d-9bea9fb65d85 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.151818] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 913.151818] env[61907]: value = "task-1244074" [ 913.151818] env[61907]: _type = "Task" [ 913.151818] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.160298] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.211440] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244072, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.396754] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244073, 'name': CreateVM_Task, 'duration_secs': 0.299901} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.396936] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.397405] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.397577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.398018] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.398296] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f9187f9-cd8a-4fdc-89ee-22b4547bb592 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.403097] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 913.403097] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a6a20-2029-ba22-0942-77f0ae7f0a6a" [ 913.403097] env[61907]: _type = "Task" [ 913.403097] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.410792] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a6a20-2029-ba22-0942-77f0ae7f0a6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.509172] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.719s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.511679] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.039s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.513797] env[61907]: INFO nova.compute.claims [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.517920] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.518358] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.518715] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.519050] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.531494] env[61907]: INFO nova.scheduler.client.report [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted allocations for instance 2c1212a7-df13-49b7-b3b3-24fb59b6789d [ 913.663102] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.184889} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.663445] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 913.664177] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cf4fa5-a6eb-4b20-86bb-a0ed45b81d15 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.682648] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] c3bf9001-7166-4064-89d5-96be073fa7c2/c3bf9001-7166-4064-89d5-96be073fa7c2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 913.682648] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b651433a-5d09-47be-9131-828dafffaded {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.701234] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 913.701234] env[61907]: value = "task-1244075" [ 913.701234] env[61907]: _type = "Task" [ 913.701234] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.713872] env[61907]: DEBUG oslo_vmware.api [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244072, 'name': PowerOnVM_Task, 'duration_secs': 0.575067} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.714106] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244075, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.714358] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 913.714560] env[61907]: INFO nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Took 7.52 seconds to spawn the instance on the hypervisor. [ 913.714737] env[61907]: DEBUG nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 913.715721] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172455b3-4fa1-4e9b-9b59-091e74367ffd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.914352] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a6a20-2029-ba22-0942-77f0ae7f0a6a, 'name': SearchDatastore_Task, 'duration_secs': 0.008307} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.915043] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.915043] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 913.915304] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.915503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.915735] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 913.916049] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3529c5af-ac6f-4eee-9bc8-5e15fa7eb3f9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.924189] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 913.924412] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 913.925221] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f07acab-7320-46d4-afc6-e2c4f3d48d70 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.930336] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 913.930336] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52496445-f97e-7940-5cd0-d5aec385ca23" [ 913.930336] env[61907]: _type = "Task" [ 913.930336] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.938189] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52496445-f97e-7940-5cd0-d5aec385ca23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.021028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.039361] env[61907]: DEBUG oslo_concurrency.lockutils [None req-517b98e0-369a-4154-8300-b2961a3e759d tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "2c1212a7-df13-49b7-b3b3-24fb59b6789d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.306s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.210716] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244075, 'name': ReconfigVM_Task, 'duration_secs': 0.293417} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.211013] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Reconfigured VM instance instance-00000051 to attach disk [datastore1] c3bf9001-7166-4064-89d5-96be073fa7c2/c3bf9001-7166-4064-89d5-96be073fa7c2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.211721] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b26353c4-1a10-4298-9052-937d435351e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.218135] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 914.218135] env[61907]: value = "task-1244076" [ 914.218135] env[61907]: _type = "Task" [ 914.218135] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.229495] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244076, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.232998] env[61907]: INFO nova.compute.manager [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Took 18.14 seconds to build instance. [ 914.442502] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52496445-f97e-7940-5cd0-d5aec385ca23, 'name': SearchDatastore_Task, 'duration_secs': 0.008394} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.443284] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c82ff11-5f3d-4c6a-8566-74cb663d438b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.448230] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 914.448230] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522f9850-fd78-db8f-fa42-0cab909c0995" [ 914.448230] env[61907]: _type = "Task" [ 914.448230] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.456710] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522f9850-fd78-db8f-fa42-0cab909c0995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.660576] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "f028beaf-4185-40d7-a730-abe91209f9d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.660915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.661202] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.661429] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.661895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.664488] env[61907]: INFO nova.compute.manager [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Terminating instance [ 914.730149] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244076, 'name': Rename_Task, 'duration_secs': 0.143269} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.730445] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 914.730699] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df458333-d8de-4b41-bb45-97fb8b568bd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.734915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-42c1a65d-c640-4161-bd5b-7cebec2d712e tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.652s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.736785] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 914.736785] env[61907]: value = "task-1244077" [ 914.736785] env[61907]: _type = "Task" [ 914.736785] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.746568] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244077, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.754629] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9105dd-4d6b-4d66-a0c2-fddf57e06a61 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.761211] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652d7f0d-9e14-402b-961b-b07c7707c131 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.791939] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ec4e17-11b8-4da4-ad18-84415e5e742d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.799160] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b208346-844e-4642-b3f9-547dbc643c33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.813773] env[61907]: DEBUG nova.compute.provider_tree [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.959037] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522f9850-fd78-db8f-fa42-0cab909c0995, 'name': SearchDatastore_Task, 'duration_secs': 0.008706} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.959201] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.959332] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.959598] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77af05e9-1036-4ef4-93bf-0e73b540de67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.966135] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 914.966135] env[61907]: value = "task-1244078" [ 914.966135] env[61907]: _type = "Task" [ 914.966135] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.973341] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244078, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.171240] env[61907]: DEBUG nova.compute.manager [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 915.171487] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 915.172497] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc7ed9c1-51ce-46c1-b987-7a147975006f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.182499] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 915.182799] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b98f6d67-cfce-4556-abe2-53eb925a97b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.189159] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19995c49-04f9-4098-87c7-a55ab7c40b8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.193312] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 915.193312] env[61907]: value = "task-1244079" [ 915.193312] env[61907]: _type = "Task" [ 915.193312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.199311] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Suspending the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 915.200047] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-eeabb55e-7762-43d9-b5fa-aaf1877e555d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.206246] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.208454] env[61907]: DEBUG oslo_vmware.api [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 915.208454] env[61907]: value = "task-1244080" [ 915.208454] env[61907]: _type = "Task" [ 915.208454] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.217311] env[61907]: DEBUG oslo_vmware.api [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244080, 'name': SuspendVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.246711] env[61907]: DEBUG oslo_vmware.api [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244077, 'name': PowerOnVM_Task, 'duration_secs': 0.415674} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.247151] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 915.247400] env[61907]: INFO nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Took 4.82 seconds to spawn the instance on the hypervisor. [ 915.247611] env[61907]: DEBUG nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 915.248539] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ca69f5-4c13-4dfd-ac31-63b81c716684 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.317223] env[61907]: DEBUG nova.scheduler.client.report [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 915.478702] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244078, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504641} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.479077] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.479358] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.479674] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea05986b-e824-4fd5-a604-a04eb485bd24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.488909] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 915.488909] env[61907]: value = "task-1244081" [ 915.488909] env[61907]: _type = "Task" [ 915.488909] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.503753] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244081, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.704295] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244079, 'name': PowerOffVM_Task, 'duration_secs': 0.458014} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.704695] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 915.704904] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 915.705264] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09a30925-671b-4764-9688-431f2aed2baa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.717610] env[61907]: DEBUG oslo_vmware.api [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244080, 'name': SuspendVM_Task} progress is 58%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.767197] env[61907]: INFO nova.compute.manager [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Took 18.08 seconds to build instance. [ 915.769537] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 915.769792] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 915.769988] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleting the datastore file [datastore1] f028beaf-4185-40d7-a730-abe91209f9d8 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 915.770288] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95ced963-94a1-4243-a5c1-a08d1ce18d60 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.776694] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 915.776694] env[61907]: value = "task-1244083" [ 915.776694] env[61907]: _type = "Task" [ 915.776694] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.786437] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244083, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.822665] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.823282] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 915.825991] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.299s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.827680] env[61907]: INFO nova.compute.claims [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.999634] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244081, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.139551} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.999937] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 916.000868] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b350370-2d58-437d-beea-04bfbd686717 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.020495] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 916.020760] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-538a3feb-5919-4e21-b71b-47c09651837f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.040147] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 916.040147] env[61907]: value = "task-1244084" [ 916.040147] env[61907]: _type = "Task" [ 916.040147] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.048382] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.219612] env[61907]: DEBUG oslo_vmware.api [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244080, 'name': SuspendVM_Task, 'duration_secs': 0.701737} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.219903] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Suspended the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 916.220107] env[61907]: DEBUG nova.compute.manager [None req-c7ef3b1a-5018-4a2a-a4af-42ecdb7ced6c tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 916.220871] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd010aed-12be-4fca-b8ed-5c3e50081fe3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.269593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b90b3901-a859-4b7d-a296-6f5c696a37fc tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.594s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.286330] env[61907]: DEBUG oslo_vmware.api [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244083, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166889} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.286607] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.286798] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 916.286977] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 916.287187] env[61907]: INFO nova.compute.manager [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 916.287432] env[61907]: DEBUG oslo.service.loopingcall [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.287632] env[61907]: DEBUG nova.compute.manager [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 916.287728] env[61907]: DEBUG nova.network.neutron [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 916.334431] env[61907]: DEBUG nova.compute.utils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.338855] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 916.338855] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.402720] env[61907]: DEBUG nova.policy [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8682003dee14f25ba212739afd3d335', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6a88771a8f84196ae271dbd57590c15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.550660] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244084, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.700989] env[61907]: DEBUG nova.compute.manager [req-7bcc5d90-d8d8-42bc-b50a-33c9e843e02b req-2331ba8a-f0cf-4f26-974b-ddadf303ba72 service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Received event network-vif-deleted-b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 916.701166] env[61907]: INFO nova.compute.manager [req-7bcc5d90-d8d8-42bc-b50a-33c9e843e02b req-2331ba8a-f0cf-4f26-974b-ddadf303ba72 service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Neutron deleted interface b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb; detaching it from the instance and deleting it from the info cache [ 916.701611] env[61907]: DEBUG nova.network.neutron [req-7bcc5d90-d8d8-42bc-b50a-33c9e843e02b req-2331ba8a-f0cf-4f26-974b-ddadf303ba72 service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.807891] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Successfully created port: 92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.838812] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 917.052305] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244084, 'name': ReconfigVM_Task, 'duration_secs': 0.732307} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.055077] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.056061] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d592903a-8497-4a82-8598-6546735ff722 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.062703] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 917.062703] env[61907]: value = "task-1244085" [ 917.062703] env[61907]: _type = "Task" [ 917.062703] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.071903] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244085, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.111298] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0120d35d-7914-4138-a5e8-8ea5a8f5d99d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.119232] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4918b017-cb04-476b-9248-df5fee1f117e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.155517] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f8a6db4-d360-4647-86ea-aa2bd743868e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.156826] env[61907]: DEBUG nova.network.neutron [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.171278] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4949abaf-7447-4b65-b3c8-7c153935daea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.186014] env[61907]: DEBUG nova.compute.provider_tree [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 917.204635] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3b6f717d-591b-4ff8-9fbd-88d7b8e3317a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.213639] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c6cc07-b491-4e58-91a7-f35cd16f3bbd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.245803] env[61907]: DEBUG nova.compute.manager [req-7bcc5d90-d8d8-42bc-b50a-33c9e843e02b req-2331ba8a-f0cf-4f26-974b-ddadf303ba72 service nova] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Detach interface failed, port_id=b2a3b1ca-0e9a-4adc-a262-7b33b3cabbcb, reason: Instance f028beaf-4185-40d7-a730-abe91209f9d8 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 917.574052] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244085, 'name': Rename_Task, 'duration_secs': 0.138059} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.574401] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 917.574663] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06ed9578-5a19-4bec-b912-604622d3caaa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.581710] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 917.581710] env[61907]: value = "task-1244086" [ 917.581710] env[61907]: _type = "Task" [ 917.581710] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.590163] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.659063] env[61907]: INFO nova.compute.manager [-] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Took 1.37 seconds to deallocate network for instance. [ 917.664614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.664980] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.665251] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.665456] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.665635] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.667695] env[61907]: INFO nova.compute.manager [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Terminating instance [ 917.689275] env[61907]: DEBUG nova.scheduler.client.report [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 917.851699] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 917.884728] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='40f1d1e6a16c03b3883d0eadd273deb1',container_format='bare',created_at=2024-10-10T13:12:54Z,direct_url=,disk_format='vmdk',id=93afda2d-a435-4b9d-983e-30d2ad798313,min_disk=1,min_ram=0,name='tempest-test-snap-458602461',owner='e6a88771a8f84196ae271dbd57590c15',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-10T13:13:10Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.885108] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.885397] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.885484] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.885665] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.885835] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.886084] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.886265] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.886446] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.886618] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.886802] env[61907]: DEBUG nova.virt.hardware [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.887797] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393f40de-acaf-4bfe-a5c6-33f8ff2d3b14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.899020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa91dbb-1e27-4072-933f-9d53fbc2b485 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.092303] env[61907]: DEBUG oslo_vmware.api [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244086, 'name': PowerOnVM_Task, 'duration_secs': 0.487125} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.092770] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 918.093015] env[61907]: INFO nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Took 5.28 seconds to spawn the instance on the hypervisor. [ 918.093216] env[61907]: DEBUG nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 918.094046] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc711a68-d3c6-4813-a85c-a939ea03a4d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.168422] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.171594] env[61907]: DEBUG nova.compute.manager [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 918.171594] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 918.172214] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcea59d1-8042-4996-a22a-e6556f7119b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.180614] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.180963] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2a82dd9f-c701-4a2d-8967-94b1d90c3e2d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.194359] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.194980] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 918.197747] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 14.334s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.248880] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.248880] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.248880] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleting the datastore file [datastore1] 4b07210f-8005-481f-b7a1-ce80a4cbe980 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.248880] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-39e29f65-cabe-43cc-8bcd-c32a049a05df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.255644] env[61907]: DEBUG oslo_vmware.api [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 918.255644] env[61907]: value = "task-1244088" [ 918.255644] env[61907]: _type = "Task" [ 918.255644] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.267526] env[61907]: DEBUG oslo_vmware.api [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.452425] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Successfully updated port: 92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.611999] env[61907]: INFO nova.compute.manager [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Took 20.72 seconds to build instance. [ 918.705436] env[61907]: DEBUG nova.compute.utils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.710359] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 918.710359] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 918.731940] env[61907]: DEBUG nova.compute.manager [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Received event network-vif-plugged-92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 918.732209] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Acquiring lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.732437] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.732614] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.732786] env[61907]: DEBUG nova.compute.manager [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] No waiting events found dispatching network-vif-plugged-92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 918.732954] env[61907]: WARNING nova.compute.manager [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Received unexpected event network-vif-plugged-92d170c7-dcdb-4764-8af5-82a44f2ba99e for instance with vm_state building and task_state spawning. [ 918.733140] env[61907]: DEBUG nova.compute.manager [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Received event network-changed-92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 918.733303] env[61907]: DEBUG nova.compute.manager [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Refreshing instance network info cache due to event network-changed-92d170c7-dcdb-4764-8af5-82a44f2ba99e. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 918.733485] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Acquiring lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.733625] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Acquired lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.733782] env[61907]: DEBUG nova.network.neutron [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Refreshing network info cache for port 92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.755723] env[61907]: DEBUG nova.policy [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.768853] env[61907]: DEBUG oslo_vmware.api [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244088, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.956115] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.970690] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5ff345-eda8-42c4-8e04-ae60e3ac2cb0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.979218] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b02c594-f481-4cde-bfc2-119f04e4146f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.018050] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ef38db4-7ee0-4c0d-bebf-3fb4d350bcea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.026796] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3e20536-8bd5-48b7-99b8-4f24fb6c200c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.045171] env[61907]: DEBUG nova.compute.provider_tree [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 919.113719] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2b89ba94-ea6e-4217-8670-40f699bbd939 tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.242s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.129257] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Successfully created port: bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.210677] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 919.267661] env[61907]: DEBUG oslo_vmware.api [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244088, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.53699} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.267914] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.268119] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 919.268311] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 919.268504] env[61907]: INFO nova.compute.manager [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Took 1.10 seconds to destroy the instance on the hypervisor. [ 919.268754] env[61907]: DEBUG oslo.service.loopingcall [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.268982] env[61907]: DEBUG nova.compute.manager [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 919.269139] env[61907]: DEBUG nova.network.neutron [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 919.278156] env[61907]: DEBUG nova.network.neutron [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.447251] env[61907]: DEBUG nova.network.neutron [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.522348] env[61907]: INFO nova.compute.manager [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Rebuilding instance [ 919.525858] env[61907]: DEBUG nova.compute.manager [req-48afa8b0-8fac-4b1e-a9ae-749feabd432e req-b2f5ddfe-f6e3-4dca-9420-f05e751b3326 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Received event network-vif-deleted-61efaa0f-fae2-4091-97b1-7937f954ff49 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 919.526082] env[61907]: INFO nova.compute.manager [req-48afa8b0-8fac-4b1e-a9ae-749feabd432e req-b2f5ddfe-f6e3-4dca-9420-f05e751b3326 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Neutron deleted interface 61efaa0f-fae2-4091-97b1-7937f954ff49; detaching it from the instance and deleting it from the info cache [ 919.526405] env[61907]: DEBUG nova.network.neutron [req-48afa8b0-8fac-4b1e-a9ae-749feabd432e req-b2f5ddfe-f6e3-4dca-9420-f05e751b3326 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.564685] env[61907]: DEBUG nova.compute.manager [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 919.565562] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc7cb803-2cbb-4690-b15f-813ef8b1481b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.578426] env[61907]: DEBUG nova.scheduler.client.report [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 100 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 919.578680] env[61907]: DEBUG nova.compute.provider_tree [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 100 to 101 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 919.578867] env[61907]: DEBUG nova.compute.provider_tree [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 919.949842] env[61907]: DEBUG oslo_concurrency.lockutils [req-a9de6f2f-7096-4550-8f70-31cd8631f58c req-c1e93d1c-9761-467c-9b19-06c8ba9c1c0c service nova] Releasing lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.950346] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.950538] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 920.002870] env[61907]: DEBUG nova.network.neutron [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.028688] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7ddfc754-f6fe-4932-9c46-78236f40fd5e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.038181] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55009642-4c89-4614-a3ea-23682bfed696 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.068237] env[61907]: DEBUG nova.compute.manager [req-48afa8b0-8fac-4b1e-a9ae-749feabd432e req-b2f5ddfe-f6e3-4dca-9420-f05e751b3326 service nova] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Detach interface failed, port_id=61efaa0f-fae2-4091-97b1-7937f954ff49, reason: Instance 4b07210f-8005-481f-b7a1-ce80a4cbe980 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 920.221011] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 920.242244] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.242522] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.242687] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.242874] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.243047] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.243236] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.243453] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.243619] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.243791] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.244014] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.244214] env[61907]: DEBUG nova.virt.hardware [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.245096] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1c5f4a-fe01-4661-bd0e-3ff67ccdfaba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.253423] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740c45b0-8acf-4acd-a23f-285813bc6b55 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.484140] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 920.505917] env[61907]: INFO nova.compute.manager [-] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Took 1.24 seconds to deallocate network for instance. [ 920.578761] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.579072] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e9b6aa0-497b-42aa-b9a6-2da1a000dcdf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.587998] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.390s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.591308] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 920.591308] env[61907]: value = "task-1244089" [ 920.591308] env[61907]: _type = "Task" [ 920.591308] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.595074] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.381s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.595074] env[61907]: INFO nova.compute.claims [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.604504] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244089, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.682094] env[61907]: DEBUG nova.network.neutron [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Updating instance_info_cache with network_info: [{"id": "92d170c7-dcdb-4764-8af5-82a44f2ba99e", "address": "fa:16:3e:b5:ca:a3", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92d170c7-dc", "ovs_interfaceid": "92d170c7-dcdb-4764-8af5-82a44f2ba99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.765714] env[61907]: DEBUG nova.compute.manager [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-vif-plugged-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 920.765929] env[61907]: DEBUG oslo_concurrency.lockutils [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.766423] env[61907]: DEBUG oslo_concurrency.lockutils [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.766619] env[61907]: DEBUG oslo_concurrency.lockutils [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.766976] env[61907]: DEBUG nova.compute.manager [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] No waiting events found dispatching network-vif-plugged-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 920.767320] env[61907]: WARNING nova.compute.manager [req-b63e85cf-5df2-4b0b-9736-c5bc5578678b req-4560d1f1-f7fd-4967-94b4-2de2d2051d30 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received unexpected event network-vif-plugged-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 for instance with vm_state building and task_state spawning. [ 920.832698] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Successfully updated port: bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.013098] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.108925] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244089, 'name': PowerOffVM_Task, 'duration_secs': 0.193951} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.109758] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.109997] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 921.110787] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9129ec26-4062-4ff9-9c39-91416cb572a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.119269] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.119500] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dc05863-f3ea-4ecc-88d5-1b7e2daff5a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.145068] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.145316] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.145502] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleting the datastore file [datastore2] 71804bc1-79ce-4731-9e27-74c007d4e906 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.146049] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3307399e-cd4c-4806-82e5-086affefb8a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.151416] env[61907]: INFO nova.scheduler.client.report [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocation for migration 50cbf55f-4d6f-40b0-af2d-1eaddbc36258 [ 921.153610] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 921.153610] env[61907]: value = "task-1244091" [ 921.153610] env[61907]: _type = "Task" [ 921.153610] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.164636] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.184910] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-23f9f3d1-b376-4502-927c-e9bd89693131" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.184910] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance network_info: |[{"id": "92d170c7-dcdb-4764-8af5-82a44f2ba99e", "address": "fa:16:3e:b5:ca:a3", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap92d170c7-dc", "ovs_interfaceid": "92d170c7-dcdb-4764-8af5-82a44f2ba99e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 921.185527] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:ca:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '92d170c7-dcdb-4764-8af5-82a44f2ba99e', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 921.192952] env[61907]: DEBUG oslo.service.loopingcall [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.193884] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 921.194147] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-632f8a72-f3fa-4d9e-a6c9-b658ae998ffe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.215236] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 921.215236] env[61907]: value = "task-1244092" [ 921.215236] env[61907]: _type = "Task" [ 921.215236] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.226388] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244092, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.335403] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.335668] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.335668] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 921.660861] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6705602-6393-4aab-a0db-7c27a5102e3f tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 21.136s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.668263] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12413} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.668397] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.668588] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.668767] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.725136] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244092, 'name': CreateVM_Task, 'duration_secs': 0.308299} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.725322] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 921.726011] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.726196] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.726833] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 921.726833] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7e6cde8-1451-4282-96b5-57b597eaf721 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.733921] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 921.733921] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a2a0e-0bf7-9aa0-e9c4-33db1561c012" [ 921.733921] env[61907]: _type = "Task" [ 921.733921] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.742487] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529a2a0e-0bf7-9aa0-e9c4-33db1561c012, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.807736] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1415510c-61f7-46ed-b04e-273eee568e56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.816151] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c36d117e-46b7-4921-9a99-bc08502054bc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.850721] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23ae2bb-8bfc-4215-92f1-2b72142f4d34 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.861691] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35a655b-0147-4c36-a6ec-5ad3f111fcee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.873363] env[61907]: DEBUG nova.compute.provider_tree [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.906398] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.075195] env[61907]: DEBUG nova.network.neutron [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.245964] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.246435] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Processing image 93afda2d-a435-4b9d-983e-30d2ad798313 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 922.246797] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.247089] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.247410] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 922.247778] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1aca73e8-a233-4468-986c-62e5541c2a99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.260675] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 922.261068] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 922.261892] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ebf84f1-22a0-45f5-adb4-689edc8475a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.270015] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 922.270015] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ddeb98-6286-c8b0-ef10-7a1599c0dd0b" [ 922.270015] env[61907]: _type = "Task" [ 922.270015] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.276950] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ddeb98-6286-c8b0-ef10-7a1599c0dd0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.376212] env[61907]: DEBUG nova.scheduler.client.report [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 922.578551] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.578894] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Instance network_info: |[{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 922.579372] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:ef:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bf950e7a-fdf3-453e-8bda-0ccca9cc8c98', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.587215] env[61907]: DEBUG oslo.service.loopingcall [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.587478] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.587729] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-093883ea-8dd3-43a6-aea5-b294ca4149a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.609211] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.609211] env[61907]: value = "task-1244093" [ 922.609211] env[61907]: _type = "Task" [ 922.609211] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.616901] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244093, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.711760] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.712194] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.712439] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.712737] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.712980] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.713232] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.713557] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.713824] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.714111] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.714370] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.714636] env[61907]: DEBUG nova.virt.hardware [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.715647] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd00e67-38de-464c-80a1-f896afe769c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.724405] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f880362-7942-41dd-ba5a-5cdd0485d328 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.738715] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.744611] env[61907]: DEBUG oslo.service.loopingcall [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.745010] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 922.745205] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b2d0d32-7982-44bb-b69f-d60a6382ae66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.764544] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.764544] env[61907]: value = "task-1244094" [ 922.764544] env[61907]: _type = "Task" [ 922.764544] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.782202] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244094, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.782623] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Preparing fetch location {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 922.782859] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Fetch image to [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4/OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4.vmdk {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 922.783097] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Downloading stream optimized image 93afda2d-a435-4b9d-983e-30d2ad798313 to [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4/OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4.vmdk on the data store datastore1 as vApp {{(pid=61907) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 922.783269] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Downloading image file data 93afda2d-a435-4b9d-983e-30d2ad798313 to the ESX as VM named 'OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4' {{(pid=61907) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 922.863123] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 922.863123] env[61907]: value = "resgroup-9" [ 922.863123] env[61907]: _type = "ResourcePool" [ 922.863123] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 922.863476] env[61907]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-492b3aee-8a9d-4cec-a8fb-f80a2b8a3f9a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.882705] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.883261] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 922.888412] env[61907]: DEBUG nova.compute.manager [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 922.888612] env[61907]: DEBUG nova.compute.manager [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 922.889340] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.889582] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.889790] env[61907]: DEBUG nova.network.neutron [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 922.892039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.871s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.892318] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.892506] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 922.892814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.725s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.893056] env[61907]: DEBUG nova.objects.instance [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'resources' on Instance uuid f028beaf-4185-40d7-a730-abe91209f9d8 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 922.895304] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbe9121-516d-41d8-aa1a-5590872e063c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.901301] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease: (returnval){ [ 922.901301] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 922.901301] env[61907]: _type = "HttpNfcLease" [ 922.901301] env[61907]: } obtained for vApp import into resource pool (val){ [ 922.901301] env[61907]: value = "resgroup-9" [ 922.901301] env[61907]: _type = "ResourcePool" [ 922.901301] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 922.901692] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the lease: (returnval){ [ 922.901692] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 922.901692] env[61907]: _type = "HttpNfcLease" [ 922.901692] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 922.910844] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0df5b1-9aed-43b1-841b-5cba858e8bfb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.918857] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 922.918857] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 922.918857] env[61907]: _type = "HttpNfcLease" [ 922.918857] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 922.930341] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fff8909-691f-4078-bcf7-43328698d4c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.941125] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fc955f-ed52-4d9d-99d9-cc2ea32d0abb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.981427] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180385MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 922.981641] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.121202] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244093, 'name': CreateVM_Task, 'duration_secs': 0.378339} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.121479] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.122311] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.122526] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.122859] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.123136] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ebbdb6-5aa9-4fb6-bf12-c1e70b0a7cc9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.128291] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 923.128291] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a8ac3b-648a-a762-7fbd-74911cf14ab9" [ 923.128291] env[61907]: _type = "Task" [ 923.128291] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.137694] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a8ac3b-648a-a762-7fbd-74911cf14ab9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.275724] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244094, 'name': CreateVM_Task, 'duration_secs': 0.293939} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.275941] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 923.276432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.390035] env[61907]: DEBUG nova.compute.utils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.391753] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 923.392039] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 923.412858] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.412858] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 923.412858] env[61907]: _type = "HttpNfcLease" [ 923.412858] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.446127] env[61907]: DEBUG nova.policy [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da98c262c7fb41dea5785c9b89f7fdb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2048329d91224454805d73742252fc4d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.451030] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.451327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.451623] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.451799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.451992] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.457695] env[61907]: INFO nova.compute.manager [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Terminating instance [ 923.557511] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.557768] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.558017] env[61907]: INFO nova.compute.manager [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Rebooting instance [ 923.627182] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3071dde2-3098-457e-ab0a-f558b6419399 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.634384] env[61907]: DEBUG nova.network.neutron [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 923.634721] env[61907]: DEBUG nova.network.neutron [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.641555] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a8ac3b-648a-a762-7fbd-74911cf14ab9, 'name': SearchDatastore_Task, 'duration_secs': 0.012853} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.642146] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.642347] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.642581] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.642735] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.642920] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 923.643960] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac156f4b-350f-4d3a-82c5-4ce1a7a71647 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.647894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.648396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.648469] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c6fef16f-657a-4615-b34f-801225a28ee6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.650354] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3f56032-f37c-4ac0-8f1d-51b9d37b3b64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.656882] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 923.656882] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb1112-3aa5-ed40-e506-4e77a446b66a" [ 923.656882] env[61907]: _type = "Task" [ 923.656882] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.690929] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5219e03e-b2fe-49f4-8723-577fe88c5d41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.694371] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 923.694927] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 923.696013] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab6bee25-a2b4-419a-856b-8fa170823cfb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.701997] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 923.701997] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e3377-b623-8541-0701-1decd43d4b84" [ 923.701997] env[61907]: _type = "Task" [ 923.701997] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.707407] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb1112-3aa5-ed40-e506-4e77a446b66a, 'name': SearchDatastore_Task, 'duration_secs': 0.01157} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.708175] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Successfully created port: c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.714686] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8435b3b1-9ad6-46d4-9e64-eda322d7f841 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.719757] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.720007] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 923.720235] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.736620] env[61907]: DEBUG nova.compute.provider_tree [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.738410] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521e3377-b623-8541-0701-1decd43d4b84, 'name': SearchDatastore_Task, 'duration_secs': 0.009919} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.739195] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ad21412-e454-4783-be61-9e55a933b8f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.745490] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 923.745490] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5219f089-b21e-7d15-2a4d-f66302f809ac" [ 923.745490] env[61907]: _type = "Task" [ 923.745490] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.755642] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5219f089-b21e-7d15-2a4d-f66302f809ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.899661] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 923.913697] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 923.913697] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 923.913697] env[61907]: _type = "HttpNfcLease" [ 923.913697] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 923.962291] env[61907]: DEBUG nova.compute.manager [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 923.962531] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.963433] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee76eea5-3003-4cae-918e-377cfce68286 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.970593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.970928] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.976564] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.977057] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d58b0ac6-01e2-4935-806b-f71b722bdf91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.985999] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 923.985999] env[61907]: value = "task-1244096" [ 923.985999] env[61907]: _type = "Task" [ 923.985999] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.995683] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.075586] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.075757] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.075993] env[61907]: DEBUG nova.network.neutron [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 924.142876] env[61907]: DEBUG oslo_concurrency.lockutils [req-9ee2a639-748b-4c11-80de-9dae33b97686 req-c37ad958-2c85-404d-b8e5-1ad8fbb07732 service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.239603] env[61907]: DEBUG nova.scheduler.client.report [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 924.257164] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5219f089-b21e-7d15-2a4d-f66302f809ac, 'name': SearchDatastore_Task, 'duration_secs': 0.010917} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.257452] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.257717] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] fcde07bf-858a-4377-b27c-4f17356306af/fcde07bf-858a-4377-b27c-4f17356306af.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.258634] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.258843] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.259089] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b492b5db-ac7e-4769-bc1b-c2b917abce98 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.261054] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f5835f92-b338-4106-934b-cbb85a55c56d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.268634] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 924.268634] env[61907]: value = "task-1244097" [ 924.268634] env[61907]: _type = "Task" [ 924.268634] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.273261] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.273450] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 924.274619] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e60d2850-5439-465b-a6c0-41d647d4685b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.279963] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244097, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.283400] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 924.283400] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523f8b43-3271-517e-1186-e276ddd7a4f1" [ 924.283400] env[61907]: _type = "Task" [ 924.283400] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.291020] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523f8b43-3271-517e-1186-e276ddd7a4f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.414808] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.414808] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 924.414808] env[61907]: _type = "HttpNfcLease" [ 924.414808] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 924.477600] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 924.496982] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244096, 'name': PowerOffVM_Task, 'duration_secs': 0.208295} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.497315] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.497513] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.497794] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c1ed136e-a1df-4fc4-b21b-033710a1c14f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.567594] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.567878] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.568134] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleting the datastore file [datastore2] 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.568445] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10a0c3a0-c8e4-4eb9-9b4c-132d3c78095f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.577515] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 924.577515] env[61907]: value = "task-1244099" [ 924.577515] env[61907]: _type = "Task" [ 924.577515] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.590852] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.745091] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.747534] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.735s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.747797] env[61907]: DEBUG nova.objects.instance [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'resources' on Instance uuid 4b07210f-8005-481f-b7a1-ce80a4cbe980 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.769529] env[61907]: INFO nova.scheduler.client.report [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted allocations for instance f028beaf-4185-40d7-a730-abe91209f9d8 [ 924.782570] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244097, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.793476] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523f8b43-3271-517e-1186-e276ddd7a4f1, 'name': SearchDatastore_Task, 'duration_secs': 0.010101} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.794435] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdc18767-8f7b-4c45-b0b8-dfa1b69e4697 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.802473] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 924.802473] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f76a1c-91c9-bcfa-28e0-fcd978820601" [ 924.802473] env[61907]: _type = "Task" [ 924.802473] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.818607] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f76a1c-91c9-bcfa-28e0-fcd978820601, 'name': SearchDatastore_Task, 'duration_secs': 0.011729} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.818761] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.819605] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 924.819605] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1137b7fd-b77c-426d-b2ea-2cad8928bdfc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.828313] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 924.828313] env[61907]: value = "task-1244100" [ 924.828313] env[61907]: _type = "Task" [ 924.828313] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.836807] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244100, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.860765] env[61907]: DEBUG nova.network.neutron [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.906356] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 924.919511] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 924.919511] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 924.919511] env[61907]: _type = "HttpNfcLease" [ 924.919511] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 924.919817] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 924.919817] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525066c9-936a-0d09-ecba-b679f67eb75e" [ 924.919817] env[61907]: _type = "HttpNfcLease" [ 924.919817] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 924.920682] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438af68c-0f3f-4b31-b3d6-8c030e0f205d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.930544] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 924.930740] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk. {{(pid=61907) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.996332] env[61907]: DEBUG nova.virt.hardware [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.997989] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e2d6877-4772-4d46-a289-d1d441b85bb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.008796] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7ab50a6f-e166-49a6-9ad4-ada5808bd747 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.016239] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-134482d6-8ec9-41f5-9d08-b2d0f8f07aab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.021465] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.091669] env[61907]: DEBUG oslo_vmware.api [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244099, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284139} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.092017] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.092231] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.092418] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.092599] env[61907]: INFO nova.compute.manager [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 925.092860] env[61907]: DEBUG oslo.service.loopingcall [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.093096] env[61907]: DEBUG nova.compute.manager [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 925.093197] env[61907]: DEBUG nova.network.neutron [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.152629] env[61907]: DEBUG nova.compute.manager [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Received event network-vif-plugged-c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 925.152947] env[61907]: DEBUG oslo_concurrency.lockutils [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] Acquiring lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.153206] env[61907]: DEBUG oslo_concurrency.lockutils [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.153248] env[61907]: DEBUG oslo_concurrency.lockutils [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.153927] env[61907]: DEBUG nova.compute.manager [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] No waiting events found dispatching network-vif-plugged-c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 925.153927] env[61907]: WARNING nova.compute.manager [req-c39b5433-7bcc-4805-8a06-8d3e9476d6e1 req-c5ceecdb-9ccd-470d-b648-4c04d66f55e7 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Received unexpected event network-vif-plugged-c59a087f-2738-4034-943b-3c6e08fd58b0 for instance with vm_state building and task_state spawning. [ 925.283417] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244097, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.533598} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.283963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c7a40c5c-9600-4c12-b828-88b9d60f93c6 tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "f028beaf-4185-40d7-a730-abe91209f9d8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.623s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.284622] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] fcde07bf-858a-4377-b27c-4f17356306af/fcde07bf-858a-4377-b27c-4f17356306af.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.284985] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.288299] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63adb786-ea47-43f0-88d2-6635d63710ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.297108] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 925.297108] env[61907]: value = "task-1244101" [ 925.297108] env[61907]: _type = "Task" [ 925.297108] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.314148] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.336427] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Successfully updated port: c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.344141] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244100, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.456163} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.345498] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 925.345498] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.345498] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-400db0d8-de1f-4045-b62e-2d82be5e56e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.356731] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 925.356731] env[61907]: value = "task-1244102" [ 925.356731] env[61907]: _type = "Task" [ 925.356731] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.368282] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.375774] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244102, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.579604] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3cb1c60-4e6d-4ca0-9c0a-37b234c93195 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.589152] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f716117-791d-4e84-a83b-8e5b724cdc92 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.636636] env[61907]: DEBUG nova.compute.manager [req-e5d9b895-c879-43e2-b375-beb870103df1 req-6334db13-649e-4a1c-b102-55cbf4ad5935 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Received event network-vif-deleted-6007e3ea-4d8d-4523-b3da-93d26643424f {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 925.636837] env[61907]: INFO nova.compute.manager [req-e5d9b895-c879-43e2-b375-beb870103df1 req-6334db13-649e-4a1c-b102-55cbf4ad5935 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Neutron deleted interface 6007e3ea-4d8d-4523-b3da-93d26643424f; detaching it from the instance and deleting it from the info cache [ 925.637165] env[61907]: DEBUG nova.network.neutron [req-e5d9b895-c879-43e2-b375-beb870103df1 req-6334db13-649e-4a1c-b102-55cbf4ad5935 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.638719] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ee0237f-2baa-4e26-a416-694e48334912 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.652984] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9d3693-c2c4-4ef4-900b-34cee8341ac5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.670467] env[61907]: DEBUG nova.compute.provider_tree [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.808764] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066483} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.810609] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.811471] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba701ed6-e5a0-4016-be0b-3dd52c94e929 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.838068] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] fcde07bf-858a-4377-b27c-4f17356306af/fcde07bf-858a-4377-b27c-4f17356306af.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.842172] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71dfbcbb-6995-401f-a168-0059783c11fc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.856681] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.856827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.856917] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.861237] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "41c680fb-5450-43f3-9acb-5218e57a7da9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.861592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.861845] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.862079] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.862295] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.864552] env[61907]: INFO nova.compute.manager [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Terminating instance [ 925.876562] env[61907]: DEBUG nova.compute.manager [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 925.877271] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bdb680-23d4-4c9f-a2ae-06f480dcc611 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.880343] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 925.880343] env[61907]: value = "task-1244103" [ 925.880343] env[61907]: _type = "Task" [ 925.880343] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.885717] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244102, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081766} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.889208] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 925.897113] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1912d7b3-f244-4c2d-9c07-c9abaa14530f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.905532] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.922113] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 925.922880] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b34ff90-35ee-41ca-8a45-849f708d1ac5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.938715] env[61907]: DEBUG nova.network.neutron [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.949257] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 925.949257] env[61907]: value = "task-1244104" [ 925.949257] env[61907]: _type = "Task" [ 925.949257] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.961197] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244104, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.027035] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Completed reading data from the image iterator. {{(pid=61907) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 926.027299] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 926.028516] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b00c5f2-64c8-4684-a3ab-660f1f15856d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.037446] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 926.037586] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 926.037841] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-f19cd66a-8cef-4d95-9c1b-5dd54687cb0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.143415] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ffcc117e-20ca-434d-bd3e-45594f4248e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.153480] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9819d754-2caa-4c37-b635-3398ee217edb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.186142] env[61907]: DEBUG nova.scheduler.client.report [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 926.189380] env[61907]: DEBUG nova.compute.manager [req-e5d9b895-c879-43e2-b375-beb870103df1 req-6334db13-649e-4a1c-b102-55cbf4ad5935 service nova] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Detach interface failed, port_id=6007e3ea-4d8d-4523-b3da-93d26643424f, reason: Instance 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 926.221013] env[61907]: DEBUG oslo_vmware.rw_handles [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/524cd445-e97e-f89f-32fe-ea3ca140ef84/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 926.221316] env[61907]: INFO nova.virt.vmwareapi.images [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Downloaded image file data 93afda2d-a435-4b9d-983e-30d2ad798313 [ 926.223087] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6d34cf-5165-4b7d-8ea2-de2ba20b7de6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.238208] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7835093-b12f-49b8-b203-06d8aeeeff5a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.298683] env[61907]: INFO nova.virt.vmwareapi.images [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] The imported VM was unregistered [ 926.301075] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Caching image {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 926.301329] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating directory with path [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 926.301612] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d71baca-770d-42e3-83b8-e528aa5a6070 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.316164] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created directory with path [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313 {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 926.316164] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4/OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4.vmdk to [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk. {{(pid=61907) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 926.316164] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-2d9c89ed-92f9-4b02-a6bd-57609dd2efa3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.322869] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 926.322869] env[61907]: value = "task-1244106" [ 926.322869] env[61907]: _type = "Task" [ 926.322869] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.331769] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.377837] env[61907]: DEBUG nova.compute.manager [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 926.378025] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.379064] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37840441-60f8-414a-b37a-0ad3073398d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.387607] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.390737] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a4e90f6-8a15-44b2-9994-bbabcc20c950 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.398100] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244103, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.399559] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 926.399559] env[61907]: value = "task-1244107" [ 926.399559] env[61907]: _type = "Task" [ 926.399559] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.404407] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.414573] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.441984] env[61907]: INFO nova.compute.manager [-] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Took 1.35 seconds to deallocate network for instance. [ 926.463453] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244104, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.572331] env[61907]: DEBUG nova.network.neutron [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Updating instance_info_cache with network_info: [{"id": "c59a087f-2738-4034-943b-3c6e08fd58b0", "address": "fa:16:3e:52:d9:ec", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc59a087f-27", "ovs_interfaceid": "c59a087f-2738-4034-943b-3c6e08fd58b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.691185] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.944s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.693680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.712s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.725664] env[61907]: INFO nova.scheduler.client.report [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance 4b07210f-8005-481f-b7a1-ce80a4cbe980 [ 926.834451] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.898801] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244103, 'name': ReconfigVM_Task, 'duration_secs': 0.88244} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.899129] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfigured VM instance instance-00000054 to attach disk [datastore1] fcde07bf-858a-4377-b27c-4f17356306af/fcde07bf-858a-4377-b27c-4f17356306af.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.899808] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a42395a5-e06b-4a93-b28b-9e9e62d57205 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.911428] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 926.911428] env[61907]: value = "task-1244108" [ 926.911428] env[61907]: _type = "Task" [ 926.911428] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.915080] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12248335-7ddb-4771-b3f6-ad2d40f95526 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.918044] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244107, 'name': PowerOffVM_Task, 'duration_secs': 0.319585} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.922946] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.923162] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 926.923869] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cbdfd665-5bf9-42eb-9794-555ee95f38f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.931372] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244108, 'name': Rename_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.933631] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Doing hard reboot of VM {{(pid=61907) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 926.933911] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-089352e5-164a-4396-93f7-cc9469d6c6e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.945864] env[61907]: DEBUG oslo_vmware.api [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 926.945864] env[61907]: value = "task-1244110" [ 926.945864] env[61907]: _type = "Task" [ 926.945864] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.950528] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.959285] env[61907]: DEBUG oslo_vmware.api [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244110, 'name': ResetVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.962973] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244104, 'name': ReconfigVM_Task, 'duration_secs': 0.800905} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.963370] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 71804bc1-79ce-4731-9e27-74c007d4e906/71804bc1-79ce-4731-9e27-74c007d4e906.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.964142] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61873874-c64a-4558-8d80-2d44379ea9b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.972325] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 926.972325] env[61907]: value = "task-1244111" [ 926.972325] env[61907]: _type = "Task" [ 926.972325] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.983589] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244111, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.010408] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.010708] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.011011] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleting the datastore file [datastore1] 41c680fb-5450-43f3-9acb-5218e57a7da9 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.011288] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-42d7da00-1fd5-438c-8c8b-6bf71ab083a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.020417] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for the task: (returnval){ [ 927.020417] env[61907]: value = "task-1244112" [ 927.020417] env[61907]: _type = "Task" [ 927.020417] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.031183] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.074930] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.075372] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Instance network_info: |[{"id": "c59a087f-2738-4034-943b-3c6e08fd58b0", "address": "fa:16:3e:52:d9:ec", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc59a087f-27", "ovs_interfaceid": "c59a087f-2738-4034-943b-3c6e08fd58b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 927.075765] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:d9:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '744515ee-aa5b-4c23-b959-b56c51da6b86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c59a087f-2738-4034-943b-3c6e08fd58b0', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 927.084089] env[61907]: DEBUG oslo.service.loopingcall [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.084462] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 927.084788] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-030184b5-6886-46ac-9418-84928ddc7dbf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.107497] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 927.107497] env[61907]: value = "task-1244113" [ 927.107497] env[61907]: _type = "Task" [ 927.107497] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.116427] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244113, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.187850] env[61907]: DEBUG nova.compute.manager [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Received event network-changed-c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 927.188135] env[61907]: DEBUG nova.compute.manager [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Refreshing instance network info cache due to event network-changed-c59a087f-2738-4034-943b-3c6e08fd58b0. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 927.188321] env[61907]: DEBUG oslo_concurrency.lockutils [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] Acquiring lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.188482] env[61907]: DEBUG oslo_concurrency.lockutils [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] Acquired lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.188642] env[61907]: DEBUG nova.network.neutron [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Refreshing network info cache for port c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 927.234692] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f5febd69-daf5-4bfd-b754-c8babdf6322b tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "4b07210f-8005-481f-b7a1-ce80a4cbe980" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.570s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.335490] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.430772] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244108, 'name': Rename_Task, 'duration_secs': 0.211097} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.432081] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.432081] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-080e94bb-e5b2-4b90-ad7f-2a61a14f5ef0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.441050] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 927.441050] env[61907]: value = "task-1244114" [ 927.441050] env[61907]: _type = "Task" [ 927.441050] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.456822] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244114, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.463627] env[61907]: DEBUG oslo_vmware.api [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244110, 'name': ResetVM_Task, 'duration_secs': 0.118128} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.464082] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Did hard reboot of VM {{(pid=61907) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 927.464347] env[61907]: DEBUG nova.compute.manager [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 927.465593] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a96b86-4a1d-4a0b-ad4c-eda7684db433 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.486431] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244111, 'name': Rename_Task, 'duration_secs': 0.285934} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.486796] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 927.487049] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bbaf594-3f10-48a3-9660-1ab0c6d1c908 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.495067] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 927.495067] env[61907]: value = "task-1244115" [ 927.495067] env[61907]: _type = "Task" [ 927.495067] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.509946] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244115, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.535337] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.618677] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244113, 'name': CreateVM_Task, 'duration_secs': 0.486025} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.618885] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 927.619699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.619880] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.620243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.620542] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e57fa57-1c3e-44ce-8d95-c0d0d35781cb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.625729] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 927.625729] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52120685-e5e7-00bc-b507-dcabce96fc6b" [ 927.625729] env[61907]: _type = "Task" [ 927.625729] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.634283] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52120685-e5e7-00bc-b507-dcabce96fc6b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.736150] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736312] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736440] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23526967-cd8f-4581-b9c5-1c270d385163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736554] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736667] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 41c680fb-5450-43f3-9acb-5218e57a7da9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736781] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance e52648ac-e068-4a63-baa0-a1c34df52197 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.736893] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737045] env[61907]: WARNING nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 927.737164] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 53176c9e-d15c-49d5-b4a9-22b780279ecb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737282] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a31110ed-c4bb-4492-95a7-0a0ef5010c28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737383] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c3bf9001-7166-4064-89d5-96be073fa7c2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737494] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 71804bc1-79ce-4731-9e27-74c007d4e906 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737601] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23f9f3d1-b376-4502-927c-e9bd89693131 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737708] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance fcde07bf-858a-4377-b27c-4f17356306af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.737812] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance cd2c70a7-a290-4d1a-91fe-5da772860603 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 927.835801] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.953323] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244114, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.984280] env[61907]: DEBUG nova.network.neutron [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Updated VIF entry in instance network info cache for port c59a087f-2738-4034-943b-3c6e08fd58b0. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 927.984662] env[61907]: DEBUG nova.network.neutron [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Updating instance_info_cache with network_info: [{"id": "c59a087f-2738-4034-943b-3c6e08fd58b0", "address": "fa:16:3e:52:d9:ec", "network": {"id": "06737145-2381-4f2b-b6bf-1d0109f716c2", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-1416171567-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2048329d91224454805d73742252fc4d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "744515ee-aa5b-4c23-b959-b56c51da6b86", "external-id": "nsx-vlan-transportzone-310", "segmentation_id": 310, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc59a087f-27", "ovs_interfaceid": "c59a087f-2738-4034-943b-3c6e08fd58b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.986073] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4933b0a2-4dcb-4201-9a29-7eabbecbf70c tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.428s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.005975] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244115, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.031813] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.038239] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.038498] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.137024] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52120685-e5e7-00bc-b507-dcabce96fc6b, 'name': SearchDatastore_Task, 'duration_secs': 0.021906} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.137373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.137631] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 928.137871] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.138040] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.138233] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 928.138511] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a5a93856-efa1-4688-b32d-1c3db04237a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.149469] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 928.149675] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 928.150487] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f3e7c0a-451b-4b80-8235-9ed5021b6b74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.157393] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 928.157393] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b1dca2-fdfe-0106-155f-55ca65a2fa50" [ 928.157393] env[61907]: _type = "Task" [ 928.157393] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.166978] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b1dca2-fdfe-0106-155f-55ca65a2fa50, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.241433] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f87e45da-187b-4aad-b7bf-b4228a8b2f1f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 928.335451] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.453049] env[61907]: DEBUG oslo_vmware.api [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244114, 'name': PowerOnVM_Task, 'duration_secs': 0.831566} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.453049] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.453263] env[61907]: INFO nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Took 8.23 seconds to spawn the instance on the hypervisor. [ 928.453449] env[61907]: DEBUG nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.454399] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23eb8b4d-f3d8-4b83-8084-6fc68d3651c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.487753] env[61907]: DEBUG oslo_concurrency.lockutils [req-1863cad9-c636-409f-9e05-36ffc38811bd req-bae38725-2330-4849-96f6-9433fc794643 service nova] Releasing lock "refresh_cache-cd2c70a7-a290-4d1a-91fe-5da772860603" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.505055] env[61907]: DEBUG oslo_vmware.api [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244115, 'name': PowerOnVM_Task, 'duration_secs': 0.779538} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.505359] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 928.505591] env[61907]: DEBUG nova.compute.manager [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 928.506434] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-519d107e-910d-4318-b606-c6a54b1d9244 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.532079] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244112, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.540835] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 928.669389] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b1dca2-fdfe-0106-155f-55ca65a2fa50, 'name': SearchDatastore_Task, 'duration_secs': 0.015093} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.670222] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-652cd2db-73e2-4332-a5bc-595889170a26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.676225] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 928.676225] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e44dbe-4a0c-cb1e-785b-3bb14b374b13" [ 928.676225] env[61907]: _type = "Task" [ 928.676225] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.684703] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e44dbe-4a0c-cb1e-785b-3bb14b374b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.687309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.687625] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.745178] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5475c612-e718-49eb-9760-9cfedbd7931d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 928.745465] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 928.745615] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3200MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 928.835658] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244106, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.500147} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.838170] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4/OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4.vmdk to [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk. [ 928.838394] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Cleaning up location [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 928.838584] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_518c8d66-f53f-4113-aae8-e95979158db4 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.839068] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a54d4964-0eaa-4398-815f-fc0c3c7b599a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.846787] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 928.846787] env[61907]: value = "task-1244116" [ 928.846787] env[61907]: _type = "Task" [ 928.846787] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.859495] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244116, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.971637] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e5bae7-1801-4a68-8d95-447c1ee7cc27 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.976718] env[61907]: INFO nova.compute.manager [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Took 25.47 seconds to build instance. [ 928.980803] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20461a9a-1dbb-41d1-b914-5927c2a3f18f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.013397] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680ff118-664d-4dd3-85a6-404a9146c17f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.024805] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f7dafa8-f95a-4d80-a557-db2edd29ffb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.028942] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.041710] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.047535] env[61907]: DEBUG oslo_vmware.api [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Task: {'id': task-1244112, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.75296} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.047942] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.048157] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.048341] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.048516] env[61907]: INFO nova.compute.manager [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Took 2.67 seconds to destroy the instance on the hypervisor. [ 929.048749] env[61907]: DEBUG oslo.service.loopingcall [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.051232] env[61907]: DEBUG nova.compute.manager [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 929.051345] env[61907]: DEBUG nova.network.neutron [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.062645] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.190022] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e44dbe-4a0c-cb1e-785b-3bb14b374b13, 'name': SearchDatastore_Task, 'duration_secs': 0.014151} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.190022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.190022] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] cd2c70a7-a290-4d1a-91fe-5da772860603/cd2c70a7-a290-4d1a-91fe-5da772860603.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.190022] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0f5397a-7e37-4843-a7d6-2bc5f4d7bacb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.191232] env[61907]: DEBUG nova.compute.utils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 929.197624] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 929.197624] env[61907]: value = "task-1244117" [ 929.197624] env[61907]: _type = "Task" [ 929.197624] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.207568] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.360490] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244116, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041055} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.360490] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.360490] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.360490] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk to [datastore1] 23f9f3d1-b376-4502-927c-e9bd89693131/23f9f3d1-b376-4502-927c-e9bd89693131.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 929.360490] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e1683b23-b850-4657-9138-ec29ae94040b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.367770] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 929.367770] env[61907]: value = "task-1244118" [ 929.367770] env[61907]: _type = "Task" [ 929.367770] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.377515] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.479344] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a40b7e53-214e-4855-8677-72ed33f998b2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.980s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.493871] env[61907]: DEBUG nova.compute.manager [req-f6ea5171-d25c-47e9-875a-6db3f3d34b9b req-2f142b47-57d2-4f65-a338-ae415074252a service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Received event network-vif-deleted-4d440707-544e-46dc-bf63-b4910c232a7a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 929.494271] env[61907]: INFO nova.compute.manager [req-f6ea5171-d25c-47e9-875a-6db3f3d34b9b req-2f142b47-57d2-4f65-a338-ae415074252a service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Neutron deleted interface 4d440707-544e-46dc-bf63-b4910c232a7a; detaching it from the instance and deleting it from the info cache [ 929.494621] env[61907]: DEBUG nova.network.neutron [req-f6ea5171-d25c-47e9-875a-6db3f3d34b9b req-2f142b47-57d2-4f65-a338-ae415074252a service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.549832] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 929.694763] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.711665] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483112} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.711838] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] cd2c70a7-a290-4d1a-91fe-5da772860603/cd2c70a7-a290-4d1a-91fe-5da772860603.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 929.711988] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 929.712307] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54d45221-3463-4314-8b96-955dd0db86e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.718819] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 929.718819] env[61907]: value = "task-1244119" [ 929.718819] env[61907]: _type = "Task" [ 929.718819] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.728399] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244119, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.756169] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "71804bc1-79ce-4731-9e27-74c007d4e906" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.756479] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.756702] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "71804bc1-79ce-4731-9e27-74c007d4e906-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.756893] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.757081] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.759380] env[61907]: INFO nova.compute.manager [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Terminating instance [ 929.828813] env[61907]: DEBUG nova.network.neutron [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.878420] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.001239] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ac9f1d0a-ddf2-4cd7-a36b-8df0b5fddf68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.016024] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fb1dce-1dc6-40a6-8cdf-2208c897be60 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.048875] env[61907]: DEBUG nova.compute.manager [req-f6ea5171-d25c-47e9-875a-6db3f3d34b9b req-2f142b47-57d2-4f65-a338-ae415074252a service nova] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Detach interface failed, port_id=4d440707-544e-46dc-bf63-b4910c232a7a, reason: Instance 41c680fb-5450-43f3-9acb-5218e57a7da9 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 930.055483] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 930.055784] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.362s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.056108] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.035s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.057794] env[61907]: INFO nova.compute.claims [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 930.229887] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244119, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104127} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.230209] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 930.231190] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4284a173-b67c-459a-acac-ca70931b5e92 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.258253] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] cd2c70a7-a290-4d1a-91fe-5da772860603/cd2c70a7-a290-4d1a-91fe-5da772860603.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 930.258253] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70d94a28-c3d1-4a7a-acc4-26bc4380f236 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.274548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "refresh_cache-71804bc1-79ce-4731-9e27-74c007d4e906" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.274548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "refresh_cache-71804bc1-79ce-4731-9e27-74c007d4e906" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.274746] env[61907]: DEBUG nova.network.neutron [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.286118] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 930.286118] env[61907]: value = "task-1244120" [ 930.286118] env[61907]: _type = "Task" [ 930.286118] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.296248] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244120, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.332176] env[61907]: INFO nova.compute.manager [-] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Took 1.28 seconds to deallocate network for instance. [ 930.380297] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.781612] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.781936] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.782207] env[61907]: INFO nova.compute.manager [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Attaching volume f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2 to /dev/sdb [ 930.798108] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244120, 'name': ReconfigVM_Task, 'duration_secs': 0.323002} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.799118] env[61907]: DEBUG nova.network.neutron [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 930.801093] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Reconfigured VM instance instance-00000055 to attach disk [datastore2] cd2c70a7-a290-4d1a-91fe-5da772860603/cd2c70a7-a290-4d1a-91fe-5da772860603.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 930.801837] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5216b7b7-b756-4e73-9991-69ec228aa8e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.809038] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 930.809038] env[61907]: value = "task-1244121" [ 930.809038] env[61907]: _type = "Task" [ 930.809038] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.817310] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244121, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.821535] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3cfac8-36b8-4672-b50f-2e20378ef3ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.828315] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11e6ff9-5869-4b8e-98d4-4262d0e6b345 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.840127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.844658] env[61907]: DEBUG nova.virt.block_device [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating existing volume attachment record: 53ff2ddd-51e9-49e4-9910-8ef151fba737 {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 930.868399] env[61907]: DEBUG nova.network.neutron [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.887187] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.298287] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.298446] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.322408] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244121, 'name': Rename_Task, 'duration_secs': 0.178229} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.322408] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.325793] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a5d856e-0fec-4856-a7d9-3e1d31cd5728 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.333835] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 931.333835] env[61907]: value = "task-1244125" [ 931.333835] env[61907]: _type = "Task" [ 931.333835] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.347307] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244125, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.374993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "refresh_cache-71804bc1-79ce-4731-9e27-74c007d4e906" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.375580] env[61907]: DEBUG nova.compute.manager [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 931.375693] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.377748] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5a2d2f-6d45-4932-a829-63e21918506b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.382277] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-239744a5-5ef1-4107-b374-23cb06d8379d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.388947] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.393671] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.395933] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41d2851f-5d56-44f6-a45d-b181338d5c8f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.398756] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7ea17f-dba0-4900-aa9e-4ff63ac927cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.402786] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.403018] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.403261] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.403460] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.403675] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.406539] env[61907]: INFO nova.compute.manager [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Terminating instance [ 931.409593] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 931.409593] env[61907]: value = "task-1244126" [ 931.409593] env[61907]: _type = "Task" [ 931.409593] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.443271] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3c3770-9454-4c8b-add2-6e31a7d9e0f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.448071] env[61907]: DEBUG nova.compute.manager [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 931.448477] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 931.449711] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa5ed7d-9889-449d-b5e0-2b0f4f5cadbb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.461093] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244126, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.466129] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.467870] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478b9929-585d-4c0c-b813-bc6987912456 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.473215] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e30b750-9bea-4177-9011-c91e1e0ef1d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.486178] env[61907]: DEBUG nova.compute.provider_tree [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 931.489185] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 931.489185] env[61907]: value = "task-1244127" [ 931.489185] env[61907]: _type = "Task" [ 931.489185] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.500797] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244127, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.526774] env[61907]: DEBUG nova.compute.manager [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 931.527086] env[61907]: DEBUG nova.compute.manager [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 931.527436] env[61907]: DEBUG oslo_concurrency.lockutils [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.527635] env[61907]: DEBUG oslo_concurrency.lockutils [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.527840] env[61907]: DEBUG nova.network.neutron [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 931.804493] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 931.847731] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244125, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.884070] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.946906] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244126, 'name': PowerOffVM_Task, 'duration_secs': 0.260957} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.947284] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.947520] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.947833] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da4fbec3-ee9f-48d3-afcb-21b018c8135e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.974816] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 931.975032] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 931.975240] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleting the datastore file [datastore1] 71804bc1-79ce-4731-9e27-74c007d4e906 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 931.975495] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44199085-2400-413c-a517-070b43fc845e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.981396] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 931.981396] env[61907]: value = "task-1244129" [ 931.981396] env[61907]: _type = "Task" [ 931.981396] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.988570] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244129, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.999122] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244127, 'name': PowerOffVM_Task, 'duration_secs': 0.236008} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.999360] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.999543] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 931.999773] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-298a193f-2ec7-4aef-9083-3feb29708822 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.009798] env[61907]: ERROR nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [req-aceba51f-8da6-4c90-83f1-af3329ad0bb0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-aceba51f-8da6-4c90-83f1-af3329ad0bb0"}]} [ 932.026367] env[61907]: DEBUG nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 932.041036] env[61907]: DEBUG nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 932.041036] env[61907]: DEBUG nova.compute.provider_tree [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.051506] env[61907]: DEBUG nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 932.054887] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.055161] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 932.071177] env[61907]: DEBUG nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 932.076531] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 932.077407] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 932.077407] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleting the datastore file [datastore2] 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 932.077407] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54966cf1-cdce-44fa-89e9-91bc2758188a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.085621] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for the task: (returnval){ [ 932.085621] env[61907]: value = "task-1244131" [ 932.085621] env[61907]: _type = "Task" [ 932.085621] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.101439] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244131, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.311988] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cc5b2c-9292-4033-a5bf-c99511e5265f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.322101] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e61db41b-86b8-4397-bc36-de9e565df724 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.326121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.355297] env[61907]: DEBUG nova.network.neutron [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 932.355645] env[61907]: DEBUG nova.network.neutron [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.360056] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa42a614-4e8f-4524-bb6d-21e3f16243a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.368181] env[61907]: DEBUG oslo_vmware.api [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244125, 'name': PowerOnVM_Task, 'duration_secs': 0.539515} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.370569] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.370872] env[61907]: INFO nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Took 7.46 seconds to spawn the instance on the hypervisor. [ 932.370971] env[61907]: DEBUG nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 932.371810] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d030c5a-32b8-4f4a-9513-82515121cd9e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.376561] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53354172-8bd5-4ba5-b3db-5cb8756c4a55 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.401109] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244118, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.536529} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.401591] env[61907]: DEBUG nova.compute.provider_tree [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.402890] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/93afda2d-a435-4b9d-983e-30d2ad798313/93afda2d-a435-4b9d-983e-30d2ad798313.vmdk to [datastore1] 23f9f3d1-b376-4502-927c-e9bd89693131/23f9f3d1-b376-4502-927c-e9bd89693131.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 932.404870] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eaaf9e-59e7-4894-8ef9-30baf89fe039 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.425752] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] 23f9f3d1-b376-4502-927c-e9bd89693131/23f9f3d1-b376-4502-927c-e9bd89693131.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 932.426607] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b68f925-fc7b-4af3-9a4e-f34a75b092ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.446943] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 932.446943] env[61907]: value = "task-1244132" [ 932.446943] env[61907]: _type = "Task" [ 932.446943] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.454339] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244132, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.490137] env[61907]: DEBUG oslo_vmware.api [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244129, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096052} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.490387] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.490575] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.490838] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.490962] env[61907]: INFO nova.compute.manager [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Took 1.12 seconds to destroy the instance on the hypervisor. [ 932.491200] env[61907]: DEBUG oslo.service.loopingcall [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.491390] env[61907]: DEBUG nova.compute.manager [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 932.491482] env[61907]: DEBUG nova.network.neutron [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 932.506486] env[61907]: DEBUG nova.network.neutron [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 932.595600] env[61907]: DEBUG oslo_vmware.api [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Task: {'id': task-1244131, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132379} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.595847] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 932.596049] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 932.596241] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 932.596490] env[61907]: INFO nova.compute.manager [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Took 1.15 seconds to destroy the instance on the hypervisor. [ 932.596734] env[61907]: DEBUG oslo.service.loopingcall [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.596928] env[61907]: DEBUG nova.compute.manager [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 932.597032] env[61907]: DEBUG nova.network.neutron [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 932.864444] env[61907]: DEBUG oslo_concurrency.lockutils [req-52014ab3-7ede-4e0a-a4d8-ba8219c30b30 req-d6d85a24-525a-4ec0-8e77-94f36948fc5e service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.913028] env[61907]: INFO nova.compute.manager [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Took 28.72 seconds to build instance. [ 932.940420] env[61907]: DEBUG nova.scheduler.client.report [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 105 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 932.940629] env[61907]: DEBUG nova.compute.provider_tree [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 105 to 106 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 932.940870] env[61907]: DEBUG nova.compute.provider_tree [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 932.957309] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.009521] env[61907]: DEBUG nova.network.neutron [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.415489] env[61907]: DEBUG oslo_concurrency.lockutils [None req-18cc96f8-60ab-4c0b-9a1d-268a24c53ae9 tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.230s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.446994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.391s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.447584] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 933.450314] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.500s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.451237] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.453344] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 4.424s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.453533] env[61907]: DEBUG nova.objects.instance [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 933.467717] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.480851] env[61907]: INFO nova.scheduler.client.report [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocations for instance 0bfdb2d0-8388-4be8-a2ee-743c029db6c3 [ 933.512745] env[61907]: INFO nova.compute.manager [-] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Took 1.02 seconds to deallocate network for instance. [ 933.527848] env[61907]: DEBUG nova.network.neutron [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.553177] env[61907]: DEBUG nova.compute.manager [req-11f8c594-7bd5-454e-b0a9-f0cb5fc85f31 req-c4e968aa-fe3d-474a-8de2-7cdc98cd1104 service nova] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Received event network-vif-deleted-40a79714-91a5-4175-94fe-d7aeae0a8ea5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 933.959056] env[61907]: DEBUG nova.compute.utils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 933.967746] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 933.968059] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 933.977662] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244132, 'name': ReconfigVM_Task, 'duration_secs': 1.140981} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.977662] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Reconfigured VM instance instance-00000053 to attach disk [datastore1] 23f9f3d1-b376-4502-927c-e9bd89693131/23f9f3d1-b376-4502-927c-e9bd89693131.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.977986] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6daaf468-e419-499e-a990-3302d907356a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.984536] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 933.984536] env[61907]: value = "task-1244134" [ 933.984536] env[61907]: _type = "Task" [ 933.984536] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.990267] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b83a3bd-5a05-4ece-bc2b-bba10bb3eee2 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "0bfdb2d0-8388-4be8-a2ee-743c029db6c3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.539s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.995811] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244134, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.026235] env[61907]: DEBUG nova.policy [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41fc76e7ed9a46a09f96f11b28556ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b57c1194e0f14ae498d99d302da346c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.028400] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.030261] env[61907]: INFO nova.compute.manager [-] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Took 1.43 seconds to deallocate network for instance. [ 934.100530] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "cd2c70a7-a290-4d1a-91fe-5da772860603" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.100839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.101106] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.101337] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.101731] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.103973] env[61907]: INFO nova.compute.manager [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Terminating instance [ 934.469217] env[61907]: DEBUG oslo_concurrency.lockutils [None req-43d591f5-483b-414e-b3df-2117e1d0ebca tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.470672] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.408s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.472198] env[61907]: INFO nova.compute.claims [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.477140] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 934.498835] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244134, 'name': Rename_Task, 'duration_secs': 0.163932} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.499537] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 934.499940] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc53d869-f7d3-4850-b5dd-fca4017411bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.506397] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 934.506397] env[61907]: value = "task-1244135" [ 934.506397] env[61907]: _type = "Task" [ 934.506397] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.515729] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244135, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.536781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.570734] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Successfully created port: 8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.607931] env[61907]: DEBUG nova.compute.manager [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 934.608261] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 934.609285] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daee7a96-8e45-4ff2-a359-488267d8b62c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.617790] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 934.618096] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e926787-38c7-4fcb-9919-966e623297c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.623741] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 934.623741] env[61907]: value = "task-1244136" [ 934.623741] env[61907]: _type = "Task" [ 934.623741] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.631188] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.016456] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244135, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.134049] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244136, 'name': PowerOffVM_Task, 'duration_secs': 0.327175} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.134473] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 935.134658] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 935.134914] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0f646572-aa51-46b3-8940-8f35087b2896 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.207948] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 935.208265] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 935.208509] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleting the datastore file [datastore2] cd2c70a7-a290-4d1a-91fe-5da772860603 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.208878] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a80e8eaf-30ce-4880-86d9-73dc892547ac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.216195] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for the task: (returnval){ [ 935.216195] env[61907]: value = "task-1244138" [ 935.216195] env[61907]: _type = "Task" [ 935.216195] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.224121] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244138, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.394419] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Volume attach. Driver type: vmdk {{(pid=61907) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 935.394685] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268293', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'name': 'volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '53176c9e-d15c-49d5-b4a9-22b780279ecb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'serial': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 935.395630] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcc0ed1-a4a9-432f-aedf-cfdb75f448eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.412192] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb69fe1-d08b-44c4-bc2a-79bf31ba4fce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.438533] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2/volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 935.438856] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60612838-cb46-4acd-b8aa-48f915e26869 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.457771] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 935.457771] env[61907]: value = "task-1244139" [ 935.457771] env[61907]: _type = "Task" [ 935.457771] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.468241] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244139, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.489495] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 935.523544] env[61907]: DEBUG oslo_vmware.api [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244135, 'name': PowerOnVM_Task, 'duration_secs': 0.632907} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.525913] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.526151] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.526320] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.526512] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.526668] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.526820] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.527044] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.527283] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.527483] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.527744] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.527990] env[61907]: DEBUG nova.virt.hardware [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.528341] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 935.528664] env[61907]: INFO nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Took 17.68 seconds to spawn the instance on the hypervisor. [ 935.528850] env[61907]: DEBUG nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 935.529616] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cebdecf-3728-4822-b776-bada2b848c50 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.534877] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b70078b-b54f-4ed0-ac7d-c2d7520c866b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.543941] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a6c4bb-218b-4af8-b3e3-dc894a1bfc5f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.728758] env[61907]: DEBUG oslo_vmware.api [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Task: {'id': task-1244138, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.199602} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.729051] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.729253] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.729436] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.729616] env[61907]: INFO nova.compute.manager [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Took 1.12 seconds to destroy the instance on the hypervisor. [ 935.729866] env[61907]: DEBUG oslo.service.loopingcall [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.730072] env[61907]: DEBUG nova.compute.manager [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 935.730169] env[61907]: DEBUG nova.network.neutron [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 935.734643] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6921170-cb08-4ec4-b36e-d9fd02d5ce3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.740450] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddaa002-cbf8-44b1-9c28-7e3ae3f9131e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.777428] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39268457-7141-4ea3-b9b6-43592cabb2ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.788946] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8b205c-133d-40bc-a22a-593aa24cb77f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.802896] env[61907]: DEBUG nova.compute.provider_tree [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.968303] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244139, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.063282] env[61907]: INFO nova.compute.manager [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Took 35.61 seconds to build instance. [ 936.066810] env[61907]: DEBUG nova.compute.manager [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Received event network-vif-plugged-8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 936.066810] env[61907]: DEBUG oslo_concurrency.lockutils [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.067272] env[61907]: DEBUG oslo_concurrency.lockutils [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.067272] env[61907]: DEBUG oslo_concurrency.lockutils [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.067471] env[61907]: DEBUG nova.compute.manager [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] No waiting events found dispatching network-vif-plugged-8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 936.067509] env[61907]: WARNING nova.compute.manager [req-cf390e46-be3d-4a8f-909c-499bc803ad54 req-9001992f-5858-455e-b099-6b133199c9b9 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Received unexpected event network-vif-plugged-8b42cc64-101e-4eec-953d-5c2cc55dd252 for instance with vm_state building and task_state spawning. [ 936.091571] env[61907]: DEBUG nova.compute.manager [req-90b3e47b-65f1-48eb-b778-7e26187f6f50 req-2ea64de8-ed06-4f13-bcc0-7b5b4c869fd3 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Received event network-vif-deleted-c59a087f-2738-4034-943b-3c6e08fd58b0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 936.091772] env[61907]: INFO nova.compute.manager [req-90b3e47b-65f1-48eb-b778-7e26187f6f50 req-2ea64de8-ed06-4f13-bcc0-7b5b4c869fd3 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Neutron deleted interface c59a087f-2738-4034-943b-3c6e08fd58b0; detaching it from the instance and deleting it from the info cache [ 936.091948] env[61907]: DEBUG nova.network.neutron [req-90b3e47b-65f1-48eb-b778-7e26187f6f50 req-2ea64de8-ed06-4f13-bcc0-7b5b4c869fd3 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.111277] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.111464] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.111616] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 936.162339] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Successfully updated port: 8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.285153] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "23f9f3d1-b376-4502-927c-e9bd89693131" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.307215] env[61907]: DEBUG nova.scheduler.client.report [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 936.469786] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244139, 'name': ReconfigVM_Task, 'duration_secs': 0.961853} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.470104] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2/volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 936.475183] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e64f2634-6c02-44ac-be46-eed45d126752 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.490354] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 936.490354] env[61907]: value = "task-1244140" [ 936.490354] env[61907]: _type = "Task" [ 936.490354] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.498190] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244140, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.555750] env[61907]: DEBUG nova.network.neutron [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.564971] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3b933e47-873e-44b2-b666-422acac84dde tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.124s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.565253] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.280s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.565487] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.565726] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.565904] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.567831] env[61907]: INFO nova.compute.manager [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Terminating instance [ 936.598023] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-257869af-3705-4433-aac9-61647b902568 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.607571] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df6e36a-d4b8-4baf-a56f-c98634be2253 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.637894] env[61907]: DEBUG nova.compute.manager [req-90b3e47b-65f1-48eb-b778-7e26187f6f50 req-2ea64de8-ed06-4f13-bcc0-7b5b4c869fd3 service nova] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Detach interface failed, port_id=c59a087f-2738-4034-943b-3c6e08fd58b0, reason: Instance cd2c70a7-a290-4d1a-91fe-5da772860603 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 936.664829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.664973] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.665284] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.812535] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.813095] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 936.815693] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.976s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.815903] env[61907]: DEBUG nova.objects.instance [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lazy-loading 'resources' on Instance uuid 41c680fb-5450-43f3-9acb-5218e57a7da9 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.001960] env[61907]: DEBUG oslo_vmware.api [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244140, 'name': ReconfigVM_Task, 'duration_secs': 0.227007} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.002323] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268293', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'name': 'volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '53176c9e-d15c-49d5-b4a9-22b780279ecb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'serial': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 937.060016] env[61907]: INFO nova.compute.manager [-] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Took 1.33 seconds to deallocate network for instance. [ 937.071594] env[61907]: DEBUG nova.compute.manager [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 937.071841] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 937.072781] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aef4a22-0698-459d-923c-62c6eff8e59b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.081306] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 937.081557] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dcb99232-7ab6-4a49-bcb1-788459220d04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.088168] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 937.088168] env[61907]: value = "task-1244141" [ 937.088168] env[61907]: _type = "Task" [ 937.088168] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.098495] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.196175] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 937.318699] env[61907]: DEBUG nova.compute.utils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 937.322432] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 937.323082] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 937.357166] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.396778] env[61907]: DEBUG nova.network.neutron [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.402027] env[61907]: DEBUG nova.policy [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4a9affafc6a4de2b9cb28671bc4c07b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd8b341077bf045e0b079bbe8aa08cae0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 937.547510] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ea5a54-57e5-4957-b034-4762122d8244 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.556168] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ad4607-92fa-41f2-8416-54ede446de26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.587091] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.588103] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a1cec6-004c-4549-99a0-66e0a7028c7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.605164] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e9a183-af34-4dd3-848b-d9d5c6f330c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.609919] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244141, 'name': PowerOffVM_Task, 'duration_secs': 0.219325} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.610312] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 937.610581] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 937.611388] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25b47910-2dee-4ff0-aac9-ff525e38a999 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.622044] env[61907]: DEBUG nova.compute.provider_tree [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.679143] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Successfully created port: e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 937.703013] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 937.703261] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 937.703456] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleting the datastore file [datastore1] 23f9f3d1-b376-4502-927c-e9bd89693131 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 937.703720] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09f0311d-0981-4a77-a178-b1450e8480fc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.711375] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 937.711375] env[61907]: value = "task-1244143" [ 937.711375] env[61907]: _type = "Task" [ 937.711375] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.720425] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244143, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.823447] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 937.862519] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.862739] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 937.862958] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.864019] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 937.864019] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 937.901352] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.901699] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Instance network_info: |[{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 937.902340] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:bb:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16c6ea68-9b0e-4ac0-a484-7a9a40533017', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b42cc64-101e-4eec-953d-5c2cc55dd252', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.909948] env[61907]: DEBUG oslo.service.loopingcall [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.910677] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.910930] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1c28ae7e-6384-4059-be6f-d01d6f8bf73e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.933171] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.933171] env[61907]: value = "task-1244144" [ 937.933171] env[61907]: _type = "Task" [ 937.933171] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.945115] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244144, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.052512] env[61907]: DEBUG nova.objects.instance [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'flavor' on Instance uuid 53176c9e-d15c-49d5-b4a9-22b780279ecb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.121115] env[61907]: DEBUG nova.compute.manager [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Received event network-changed-8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 938.121335] env[61907]: DEBUG nova.compute.manager [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Refreshing instance network info cache due to event network-changed-8b42cc64-101e-4eec-953d-5c2cc55dd252. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 938.121563] env[61907]: DEBUG oslo_concurrency.lockutils [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] Acquiring lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.121714] env[61907]: DEBUG oslo_concurrency.lockutils [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] Acquired lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.121883] env[61907]: DEBUG nova.network.neutron [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Refreshing network info cache for port 8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 938.125385] env[61907]: DEBUG nova.scheduler.client.report [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 938.198598] env[61907]: INFO nova.compute.manager [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Rescuing [ 938.198855] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.199025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.199208] env[61907]: DEBUG nova.network.neutron [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 938.250082] env[61907]: DEBUG oslo_vmware.api [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244143, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139529} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.250934] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 938.251240] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 938.251489] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 938.251772] env[61907]: INFO nova.compute.manager [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Took 1.18 seconds to destroy the instance on the hypervisor. [ 938.252131] env[61907]: DEBUG oslo.service.loopingcall [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.252409] env[61907]: DEBUG nova.compute.manager [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 938.252530] env[61907]: DEBUG nova.network.neutron [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 938.444650] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244144, 'name': CreateVM_Task, 'duration_secs': 0.422566} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.444815] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.445526] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.445704] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.446036] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.446292] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-345655b4-0139-4890-8aaf-eb20fe4a68a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.450591] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 938.450591] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52885d54-ccd7-fe8f-8555-bbbbbe0b8364" [ 938.450591] env[61907]: _type = "Task" [ 938.450591] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.457761] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52885d54-ccd7-fe8f-8555-bbbbbe0b8364, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.517597] env[61907]: DEBUG nova.compute.manager [req-a3e5d23b-16e9-43d3-b9d2-6440fee7aeac req-04855d9d-4b8c-4a42-999f-0768e86b8e8f service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Received event network-vif-deleted-92d170c7-dcdb-4764-8af5-82a44f2ba99e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 938.517888] env[61907]: INFO nova.compute.manager [req-a3e5d23b-16e9-43d3-b9d2-6440fee7aeac req-04855d9d-4b8c-4a42-999f-0768e86b8e8f service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Neutron deleted interface 92d170c7-dcdb-4764-8af5-82a44f2ba99e; detaching it from the instance and deleting it from the info cache [ 938.517983] env[61907]: DEBUG nova.network.neutron [req-a3e5d23b-16e9-43d3-b9d2-6440fee7aeac req-04855d9d-4b8c-4a42-999f-0768e86b8e8f service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.557806] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6da41ce4-a454-46cb-91b8-a4f47250b85d tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.775s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.629389] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.631365] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.305s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.632891] env[61907]: INFO nova.compute.claims [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 938.652545] env[61907]: INFO nova.scheduler.client.report [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Deleted allocations for instance 41c680fb-5450-43f3-9acb-5218e57a7da9 [ 938.836663] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 938.864574] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.864826] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.864988] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.865192] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.865360] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.865514] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.865723] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.865894] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.866248] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.866475] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.866693] env[61907]: DEBUG nova.virt.hardware [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.867612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dfcdc6-dc12-4859-a96d-9e587c82bfd3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.875852] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d96970d-9e75-4619-af66-a498c2a248cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.937180] env[61907]: DEBUG nova.network.neutron [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.961870] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52885d54-ccd7-fe8f-8555-bbbbbe0b8364, 'name': SearchDatastore_Task, 'duration_secs': 0.008571} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.963639] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.964523] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.965327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.965549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.968278] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.968811] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76c77944-d684-4dce-a96e-be853b488018 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.977233] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.977488] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.978275] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c39631ca-d2a7-4af0-9ff5-18feaa908b24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.983419] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 938.983419] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c562a-6ee7-fe2b-19ec-3cfd4a1b7a10" [ 938.983419] env[61907]: _type = "Task" [ 938.983419] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.992399] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c562a-6ee7-fe2b-19ec-3cfd4a1b7a10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.994596] env[61907]: DEBUG nova.network.neutron [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.020790] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d0b1477-4d0d-46a4-a311-a22d88eded6e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.030217] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d67a04-7a51-4a51-9600-4fdcc54c6dc9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.063477] env[61907]: DEBUG nova.compute.manager [req-a3e5d23b-16e9-43d3-b9d2-6440fee7aeac req-04855d9d-4b8c-4a42-999f-0768e86b8e8f service nova] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Detach interface failed, port_id=92d170c7-dcdb-4764-8af5-82a44f2ba99e, reason: Instance 23f9f3d1-b376-4502-927c-e9bd89693131 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 939.127119] env[61907]: DEBUG nova.network.neutron [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updated VIF entry in instance network info cache for port 8b42cc64-101e-4eec-953d-5c2cc55dd252. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 939.127119] env[61907]: DEBUG nova.network.neutron [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.161884] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dbd950ae-c4b9-49f5-83c7-dccba51c162c tempest-ListServerFiltersTestJSON-1788917906 tempest-ListServerFiltersTestJSON-1788917906-project-member] Lock "41c680fb-5450-43f3-9acb-5218e57a7da9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.300s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.344460] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Successfully updated port: e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 939.439984] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.494889] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c562a-6ee7-fe2b-19ec-3cfd4a1b7a10, 'name': SearchDatastore_Task, 'duration_secs': 0.008579} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.496111] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be9ba125-0066-42a2-86d4-c67a4050c80b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.498622] env[61907]: INFO nova.compute.manager [-] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Took 1.25 seconds to deallocate network for instance. [ 939.508198] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 939.508198] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527a2e60-ccdb-ae25-df2c-2504af91afc1" [ 939.508198] env[61907]: _type = "Task" [ 939.508198] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.517887] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527a2e60-ccdb-ae25-df2c-2504af91afc1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.630029] env[61907]: DEBUG oslo_concurrency.lockutils [req-793f0a47-6815-4e0d-9714-af80d317ced7 req-3617c815-c653-47f8-8185-c1ebc14af799 service nova] Releasing lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.837711] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d430fd51-cfff-4ece-a399-82d4f90532dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.845799] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e21864-d30b-4d0e-acaf-1a06a8bc321a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.848947] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.849098] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.849286] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.878315] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c0eaafc-4fa1-4dbb-a7ef-fbeca8b820b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.885743] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e49ed3-ec09-43ff-81cd-5ec2601a1de0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.899321] env[61907]: DEBUG nova.compute.provider_tree [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.009702] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.019590] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527a2e60-ccdb-ae25-df2c-2504af91afc1, 'name': SearchDatastore_Task, 'duration_secs': 0.01546} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.019860] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.020139] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 940.020393] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-053a5d1b-120b-45f1-8d53-1ab52ccd7c2a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.027028] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 940.027028] env[61907]: value = "task-1244145" [ 940.027028] env[61907]: _type = "Task" [ 940.027028] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.034484] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244145, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.387474] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 940.402203] env[61907]: DEBUG nova.scheduler.client.report [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 940.472654] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.473000] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a75d798-0845-4bb9-9b54-3be258bdd0d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.479895] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 940.479895] env[61907]: value = "task-1244146" [ 940.479895] env[61907]: _type = "Task" [ 940.479895] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.488887] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244146, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.537654] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244145, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447749} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.537952] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 940.538221] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.538489] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b650f995-2447-4bdd-a222-209591572b8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.544193] env[61907]: DEBUG nova.compute.manager [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Received event network-vif-plugged-e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 940.544436] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.544659] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.544836] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.545016] env[61907]: DEBUG nova.compute.manager [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] No waiting events found dispatching network-vif-plugged-e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 940.545202] env[61907]: WARNING nova.compute.manager [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Received unexpected event network-vif-plugged-e44f586e-da28-4e87-95ea-518bbf489218 for instance with vm_state building and task_state spawning. [ 940.545557] env[61907]: DEBUG nova.compute.manager [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Received event network-changed-e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 940.545771] env[61907]: DEBUG nova.compute.manager [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Refreshing instance network info cache due to event network-changed-e44f586e-da28-4e87-95ea-518bbf489218. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 940.545949] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Acquiring lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.547753] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 940.547753] env[61907]: value = "task-1244147" [ 940.547753] env[61907]: _type = "Task" [ 940.547753] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.556844] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244147, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.624674] env[61907]: DEBUG nova.network.neutron [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.909651] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.278s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.910222] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 940.913475] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.885s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.913475] env[61907]: DEBUG nova.objects.instance [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lazy-loading 'resources' on Instance uuid 71804bc1-79ce-4731-9e27-74c007d4e906 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.991987] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244146, 'name': PowerOffVM_Task, 'duration_secs': 0.175518} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.992365] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 940.993162] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29665c7e-36f4-482b-b26e-e7a7123d9846 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.013913] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00e0ab2-4e6d-402b-b0f2-329b43eaebaa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.045965] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.046285] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77d86343-00e2-434b-bcb9-56148dabe9d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.053216] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 941.053216] env[61907]: value = "task-1244148" [ 941.053216] env[61907]: _type = "Task" [ 941.053216] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.056037] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244147, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060918} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.059041] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.059772] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3128c72d-87d9-4e87-88c9-25c58299cf91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.076226] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 941.076440] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.076695] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.076853] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.077050] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.086137] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.086418] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b1096a9-fac2-4ed7-bb72-07711328c7c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.088481] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a666b9e1-1679-40ef-a484-46a619982cba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.108582] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 941.108582] env[61907]: value = "task-1244149" [ 941.108582] env[61907]: _type = "Task" [ 941.108582] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.112589] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.112770] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.113786] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69c9e117-a077-488e-8712-cdb88862fc23 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.119039] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244149, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.121838] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 941.121838] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521c4594-fe22-37f6-7292-543f3ddc206f" [ 941.121838] env[61907]: _type = "Task" [ 941.121838] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.128948] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.129289] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Instance network_info: |[{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 941.129558] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521c4594-fe22-37f6-7292-543f3ddc206f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.129865] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Acquired lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.129978] env[61907]: DEBUG nova.network.neutron [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Refreshing network info cache for port e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 941.131147] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:d4:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1fb81f98-6f5a-47ab-a512-27277591d064', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e44f586e-da28-4e87-95ea-518bbf489218', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 941.138463] env[61907]: DEBUG oslo.service.loopingcall [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 941.141246] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 941.141724] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1192b7ab-65cb-4a03-92ac-60ed0f94ebc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.161863] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 941.161863] env[61907]: value = "task-1244150" [ 941.161863] env[61907]: _type = "Task" [ 941.161863] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.168306] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244150, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.417214] env[61907]: DEBUG nova.compute.utils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 941.418640] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 941.418855] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 941.434485] env[61907]: DEBUG nova.network.neutron [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updated VIF entry in instance network info cache for port e44f586e-da28-4e87-95ea-518bbf489218. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.434843] env[61907]: DEBUG nova.network.neutron [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.514725] env[61907]: DEBUG nova.policy [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 941.619941] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244149, 'name': ReconfigVM_Task, 'duration_secs': 0.445697} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.620379] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfigured VM instance instance-00000056 to attach disk [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 941.621033] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45e48cc0-5964-4328-9a98-bcc0effb91f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.630536] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 941.630536] env[61907]: value = "task-1244151" [ 941.630536] env[61907]: _type = "Task" [ 941.630536] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.638368] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]521c4594-fe22-37f6-7292-543f3ddc206f, 'name': SearchDatastore_Task, 'duration_secs': 0.00967} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.642293] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e8e98d2-b3b0-4c05-8ece-62a70d30ce45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.647673] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244151, 'name': Rename_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.650922] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 941.650922] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cef196-aaf0-e3bf-e8d3-5778182f4255" [ 941.650922] env[61907]: _type = "Task" [ 941.650922] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.660836] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cef196-aaf0-e3bf-e8d3-5778182f4255, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.668525] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244150, 'name': CreateVM_Task, 'duration_secs': 0.484661} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.668685] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 941.669396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.669548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.669901] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.671087] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23835f3f-9db6-4891-adaf-8f5a6b8e7b2d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.673114] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a0178c9-9034-4629-b097-85a9944eda3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.678371] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 941.678371] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5283e1fd-7206-3191-eb37-770f5ac495d2" [ 941.678371] env[61907]: _type = "Task" [ 941.678371] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.684172] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8ce25e-f82c-4190-a6be-4c3b8c648752 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.692509] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5283e1fd-7206-3191-eb37-770f5ac495d2, 'name': SearchDatastore_Task, 'duration_secs': 0.008126} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.716307] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.716579] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.716833] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.716979] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.717180] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.717646] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a0f8f62-d0cb-471b-9822-4d20f9771f3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.719972] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95bd22f1-bcb1-4b67-9dde-47559a9bfea9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.730075] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4d0657-c6d0-40bb-9706-97d7964ad625 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.733905] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.734100] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.734918] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7764d50-13e9-4d78-a39e-67facacd0ea1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.747435] env[61907]: DEBUG nova.compute.provider_tree [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.749886] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 941.749886] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526f3ea9-ea47-f015-3fd1-82ad60a64ce0" [ 941.749886] env[61907]: _type = "Task" [ 941.749886] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.758302] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526f3ea9-ea47-f015-3fd1-82ad60a64ce0, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.759126] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e994106a-3e52-4a8f-9e63-e6ffaef74d84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.763973] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 941.763973] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52133763-2be0-a6a7-5c9e-7f129c3c545d" [ 941.763973] env[61907]: _type = "Task" [ 941.763973] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.771365] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52133763-2be0-a6a7-5c9e-7f129c3c545d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.840146] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Successfully created port: 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 941.930693] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 941.937290] env[61907]: DEBUG oslo_concurrency.lockutils [req-00ad4b9b-bfb6-4614-b4c6-07b4cee1d112 req-c868a0a9-f8e5-4dea-a7c3-3aeb6b9ef280 service nova] Releasing lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.969200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.969534] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.969800] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.970051] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.970276] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.972922] env[61907]: INFO nova.compute.manager [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Terminating instance [ 942.140678] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244151, 'name': Rename_Task, 'duration_secs': 0.274174} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.140972] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.141233] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6bf0d2a-7965-4e6b-9772-8438237fe07d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.146643] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 942.146643] env[61907]: value = "task-1244152" [ 942.146643] env[61907]: _type = "Task" [ 942.146643] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.153523] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244152, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.160717] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cef196-aaf0-e3bf-e8d3-5778182f4255, 'name': SearchDatastore_Task, 'duration_secs': 0.010377} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.160964] env[61907]: DEBUG oslo_concurrency.lockutils [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.161239] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 942.161547] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c004a68f-3c65-4721-86c8-338339359e12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.167488] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 942.167488] env[61907]: value = "task-1244153" [ 942.167488] env[61907]: _type = "Task" [ 942.167488] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.174122] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244153, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.251947] env[61907]: DEBUG nova.scheduler.client.report [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 942.274940] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52133763-2be0-a6a7-5c9e-7f129c3c545d, 'name': SearchDatastore_Task, 'duration_secs': 0.00826} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.275221] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.275519] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.275789] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-beb8346b-a053-4d13-925e-8f81844a30c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.281721] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 942.281721] env[61907]: value = "task-1244154" [ 942.281721] env[61907]: _type = "Task" [ 942.281721] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.289159] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.477538] env[61907]: DEBUG nova.compute.manager [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 942.477743] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 942.479388] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7396cf9b-23bc-4b61-bea4-537a837dcb24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.488778] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 942.489087] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-39af2671-6d56-41d8-a783-958f39788423 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.497375] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 942.497375] env[61907]: value = "task-1244155" [ 942.497375] env[61907]: _type = "Task" [ 942.497375] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.507468] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244155, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.659345] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244152, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.685034] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244153, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453077} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.685034] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 942.685586] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4095d9df-5414-4297-8f79-0291cda75ead {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.721368] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 942.721837] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-170ae638-d2bd-477d-ac03-7538c159d3f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.743428] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 942.743428] env[61907]: value = "task-1244156" [ 942.743428] env[61907]: _type = "Task" [ 942.743428] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.752509] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244156, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.757558] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.760172] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.224s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.760438] env[61907]: DEBUG nova.objects.instance [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lazy-loading 'resources' on Instance uuid 691fff4a-9dbd-47f2-8908-69057a2bb6fe {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.787159] env[61907]: INFO nova.scheduler.client.report [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleted allocations for instance 71804bc1-79ce-4731-9e27-74c007d4e906 [ 942.794588] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244154, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.941960] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 942.968993] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.969292] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.969458] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.969676] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.969829] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.969985] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.970237] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.970475] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.970663] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.970871] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.971076] env[61907]: DEBUG nova.virt.hardware [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.972322] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1aa5469-6ab4-442f-b559-9a4bcc673f95 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.980625] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3e3c13-c687-41e3-9741-d7fea96156b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.005735] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244155, 'name': PowerOffVM_Task, 'duration_secs': 0.208204} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.005735] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 943.005735] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 943.005947] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a527ecc1-48f5-48d4-9b4a-fd77b65662f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.078355] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 943.078595] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 943.078819] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Deleting the datastore file [datastore2] a31110ed-c4bb-4492-95a7-0a0ef5010c28 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.079112] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80b7c2e3-b613-4708-8e93-d9274ffb123f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.085123] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for the task: (returnval){ [ 943.085123] env[61907]: value = "task-1244158" [ 943.085123] env[61907]: _type = "Task" [ 943.085123] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.094458] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244158, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.157501] env[61907]: DEBUG oslo_vmware.api [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244152, 'name': PowerOnVM_Task, 'duration_secs': 0.535847} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.157501] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 943.157661] env[61907]: INFO nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Took 7.67 seconds to spawn the instance on the hypervisor. [ 943.157842] env[61907]: DEBUG nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 943.158616] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7e9836-1374-48ef-9871-2eb3eb19f174 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.253229] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.266233] env[61907]: DEBUG nova.compute.manager [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-vif-plugged-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 943.266233] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.266233] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.266233] env[61907]: DEBUG oslo_concurrency.lockutils [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.266233] env[61907]: DEBUG nova.compute.manager [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] No waiting events found dispatching network-vif-plugged-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 943.266233] env[61907]: WARNING nova.compute.manager [req-5f0bf630-906f-4642-ab19-46cbb6ac9df0 req-1aa9f5e9-9035-4392-92af-f0e920d17954 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received unexpected event network-vif-plugged-0662d339-0c3c-49e2-9c37-6dd66f7e704e for instance with vm_state building and task_state spawning. [ 943.295248] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244154, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.657191} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.297673] env[61907]: DEBUG oslo_concurrency.lockutils [None req-09493c90-023e-4ec5-b610-855111bd3dcb tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "71804bc1-79ce-4731-9e27-74c007d4e906" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.541s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.298588] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 943.298863] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.299448] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a6566d8b-2b49-42a7-a0be-2ecbd49af519 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.306586] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 943.306586] env[61907]: value = "task-1244159" [ 943.306586] env[61907]: _type = "Task" [ 943.306586] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.320094] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244159, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.350694] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Successfully updated port: 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 943.466576] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a4a3d25-6424-4b7f-ae4f-8e7f9205f850 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.476048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eefd973a-2815-46e1-8b40-732beb78644f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.511058] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c150da85-6adc-449d-a0a4-a0188dff31c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.522830] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dba63ef-8787-478d-851f-3029d7be9c04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.537449] env[61907]: DEBUG nova.compute.provider_tree [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.596242] env[61907]: DEBUG oslo_vmware.api [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Task: {'id': task-1244158, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13648} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.596498] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.596828] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 943.597047] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 943.597237] env[61907]: INFO nova.compute.manager [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Took 1.12 seconds to destroy the instance on the hypervisor. [ 943.597483] env[61907]: DEBUG oslo.service.loopingcall [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.597684] env[61907]: DEBUG nova.compute.manager [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 943.597781] env[61907]: DEBUG nova.network.neutron [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 943.676641] env[61907]: INFO nova.compute.manager [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Took 18.69 seconds to build instance. [ 943.755086] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244156, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.816579] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244159, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06834} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.816878] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.817649] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0b87d7e-0579-4af3-8d95-4159a73a308c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.839668] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.839934] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d8fa71f2-f572-4eef-976a-91315ee28535 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.854007] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.854163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.854311] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 943.861106] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 943.861106] env[61907]: value = "task-1244160" [ 943.861106] env[61907]: _type = "Task" [ 943.861106] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.868650] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244160, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.017843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "c3bf9001-7166-4064-89d5-96be073fa7c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.018168] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.018395] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "c3bf9001-7166-4064-89d5-96be073fa7c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.018619] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.018827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.020969] env[61907]: INFO nova.compute.manager [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Terminating instance [ 944.042195] env[61907]: DEBUG nova.scheduler.client.report [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 944.180869] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0bd84ba9-0a32-43e2-bad1-c867e270d6da tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.210s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.255226] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244156, 'name': ReconfigVM_Task, 'duration_secs': 1.18521} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.255567] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.256463] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a53a35-2e6c-411b-a1bd-25a221b2c573 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.288676] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-629f0fd5-0239-4356-a741-60baf512f4b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.305980] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 944.305980] env[61907]: value = "task-1244161" [ 944.305980] env[61907]: _type = "Task" [ 944.305980] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.314856] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244161, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.371290] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244160, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.385991] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 944.527736] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "refresh_cache-c3bf9001-7166-4064-89d5-96be073fa7c2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.527950] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquired lock "refresh_cache-c3bf9001-7166-4064-89d5-96be073fa7c2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.528143] env[61907]: DEBUG nova.network.neutron [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.539385] env[61907]: DEBUG nova.network.neutron [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.545363] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.785s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.548737] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.962s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.549043] env[61907]: DEBUG nova.objects.instance [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lazy-loading 'resources' on Instance uuid cd2c70a7-a290-4d1a-91fe-5da772860603 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.563735] env[61907]: INFO nova.scheduler.client.report [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Deleted allocations for instance 691fff4a-9dbd-47f2-8908-69057a2bb6fe [ 944.667943] env[61907]: DEBUG nova.network.neutron [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.671184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.671184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.817480] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244161, 'name': ReconfigVM_Task, 'duration_secs': 0.378737} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.817775] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.818038] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a616de36-3a18-4e58-8a9f-b9f7cbc9266a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.824713] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 944.824713] env[61907]: value = "task-1244162" [ 944.824713] env[61907]: _type = "Task" [ 944.824713] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.831838] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244162, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.870520] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244160, 'name': ReconfigVM_Task, 'duration_secs': 0.882699} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.870827] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.871473] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e99e0b5-7568-42a4-bb53-8750e928a92e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.877751] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 944.877751] env[61907]: value = "task-1244163" [ 944.877751] env[61907]: _type = "Task" [ 944.877751] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.885658] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244163, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.041627] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.041969] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Instance network_info: |[{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 945.042345] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:63:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0662d339-0c3c-49e2-9c37-6dd66f7e704e', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 945.050444] env[61907]: DEBUG oslo.service.loopingcall [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 945.051329] env[61907]: DEBUG nova.network.neutron [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.053120] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 945.055706] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-30b3f01d-0e82-418b-b7ed-9b796cb236e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.075098] env[61907]: DEBUG oslo_concurrency.lockutils [None req-82809537-e200-4b5d-aeeb-13b680bab491 tempest-AttachVolumeShelveTestJSON-833581797 tempest-AttachVolumeShelveTestJSON-833581797-project-member] Lock "691fff4a-9dbd-47f2-8908-69057a2bb6fe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.672s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.079776] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 945.079776] env[61907]: value = "task-1244164" [ 945.079776] env[61907]: _type = "Task" [ 945.079776] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.091473] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244164, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.122870] env[61907]: DEBUG nova.network.neutron [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.171890] env[61907]: INFO nova.compute.manager [-] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Took 1.57 seconds to deallocate network for instance. [ 945.172697] env[61907]: INFO nova.compute.manager [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Detaching volume c4e7f5a7-d519-4e29-b4c1-21dcb964d875 [ 945.219779] env[61907]: INFO nova.virt.block_device [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Attempting to driver detach volume c4e7f5a7-d519-4e29-b4c1-21dcb964d875 from mountpoint /dev/sdb [ 945.220043] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Volume detach. Driver type: vmdk {{(pid=61907) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 945.220686] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268278', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'name': 'volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4c70ef23-fd26-4e90-b227-13aa050ff46d', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'serial': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 945.221404] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b0c967-adbb-4d20-a2b5-d588fe421a9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.250858] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d44cfd-a1d1-4949-98a2-8b6110a74f88 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.259501] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac69665-59ae-4455-a3ae-30993839a86c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.282819] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361576c8-8644-40ca-8bb7-78c57141f749 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.302363] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] The volume has not been displaced from its original location: [datastore2] volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875/volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875.vmdk. No consolidation needed. {{(pid=61907) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 945.308339] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 945.312835] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-413b15ea-40b8-42ef-aaeb-bc6a477b2c08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.330742] env[61907]: DEBUG nova.compute.manager [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 945.330955] env[61907]: DEBUG nova.compute.manager [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing instance network info cache due to event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 945.331269] env[61907]: DEBUG oslo_concurrency.lockutils [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.331474] env[61907]: DEBUG oslo_concurrency.lockutils [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.331806] env[61907]: DEBUG nova.network.neutron [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 945.345025] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244162, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.346678] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 945.346678] env[61907]: value = "task-1244165" [ 945.346678] env[61907]: _type = "Task" [ 945.346678] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.348329] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2750baec-88f4-4eb3-b657-c778a5bbbd9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.362768] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b4351c5-53fb-4671-b865-7e3600624c9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.366624] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.401192] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdf9205-a19c-4372-b0fb-a26b33ce70a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.410582] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244163, 'name': Rename_Task, 'duration_secs': 0.175187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.415062] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 945.415700] env[61907]: DEBUG nova.compute.manager [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 945.418746] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec0f9e1a-87cb-485c-a4da-2b238f9858f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.421704] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c320af-13f7-4c75-b908-0e2ed09b8631 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.438412] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 945.438412] env[61907]: value = "task-1244166" [ 945.438412] env[61907]: _type = "Task" [ 945.438412] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.438897] env[61907]: DEBUG nova.compute.provider_tree [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.448827] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244166, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.588985] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244164, 'name': CreateVM_Task, 'duration_secs': 0.356749} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.589165] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 945.589850] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.590218] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.590369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 945.590627] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b122f01a-d30f-4703-90c8-7ef2f854388d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.597154] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 945.597154] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ee2799-43db-487a-81a1-5676eec51a9e" [ 945.597154] env[61907]: _type = "Task" [ 945.597154] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.606042] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ee2799-43db-487a-81a1-5676eec51a9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.625406] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Releasing lock "refresh_cache-c3bf9001-7166-4064-89d5-96be073fa7c2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.626089] env[61907]: DEBUG nova.compute.manager [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 945.626921] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 945.627245] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b071cd3-5391-4fee-8007-6f9e94f07252 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.634206] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.634492] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac473024-6962-4b2c-bb71-203817664873 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.641019] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 945.641019] env[61907]: value = "task-1244167" [ 945.641019] env[61907]: _type = "Task" [ 945.641019] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.648906] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244167, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.682434] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.845352] env[61907]: DEBUG oslo_vmware.api [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244162, 'name': PowerOnVM_Task, 'duration_secs': 0.526678} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.845657] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.849238] env[61907]: DEBUG nova.compute.manager [None req-52fdad7e-74ae-4f0d-abd6-2af7bc978daf tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 945.850919] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c376b050-9647-45ad-8fcd-65b6d5efde89 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.866270] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244165, 'name': ReconfigVM_Task, 'duration_secs': 0.268772} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.866270] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 945.874864] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5def06e-f1af-480d-a052-700bf4800f11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.894370] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 945.894370] env[61907]: value = "task-1244168" [ 945.894370] env[61907]: _type = "Task" [ 945.894370] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.902035] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244168, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.945812] env[61907]: DEBUG nova.scheduler.client.report [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 945.949811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.957888] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244166, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.092637] env[61907]: DEBUG nova.network.neutron [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updated VIF entry in instance network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 946.093055] env[61907]: DEBUG nova.network.neutron [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.107688] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ee2799-43db-487a-81a1-5676eec51a9e, 'name': SearchDatastore_Task, 'duration_secs': 0.01031} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.108028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.108273] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 946.108510] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.108663] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.108848] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 946.109130] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd30433b-0be2-44a7-8254-d5b09b4a5d94 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.123262] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 946.123423] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 946.124133] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0801f8bb-ab48-4105-887a-226569c59a6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.129343] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 946.129343] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52819399-5d4f-9008-b391-195d66ec3a49" [ 946.129343] env[61907]: _type = "Task" [ 946.129343] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.138947] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52819399-5d4f-9008-b391-195d66ec3a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.149457] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244167, 'name': PowerOffVM_Task, 'duration_secs': 0.182415} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.149722] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.149896] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.150168] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-257411dd-4ba3-43f8-b1fb-ee593289de7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.172434] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.172757] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.172947] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleting the datastore file [datastore1] c3bf9001-7166-4064-89d5-96be073fa7c2 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.173219] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a1ca19de-8d5b-40f0-a307-ed505b4c0b13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.181651] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for the task: (returnval){ [ 946.181651] env[61907]: value = "task-1244170" [ 946.181651] env[61907]: _type = "Task" [ 946.181651] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.188925] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244170, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.401623] env[61907]: DEBUG oslo_vmware.api [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244168, 'name': ReconfigVM_Task, 'duration_secs': 0.187676} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.401941] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268278', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'name': 'volume-c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4c70ef23-fd26-4e90-b227-13aa050ff46d', 'attached_at': '', 'detached_at': '', 'volume_id': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875', 'serial': 'c4e7f5a7-d519-4e29-b4c1-21dcb964d875'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 946.450461] env[61907]: DEBUG oslo_vmware.api [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244166, 'name': PowerOnVM_Task, 'duration_secs': 0.73537} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.451047] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 946.451249] env[61907]: INFO nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Took 7.61 seconds to spawn the instance on the hypervisor. [ 946.451436] env[61907]: DEBUG nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 946.452163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.455154] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be0f64b2-616c-40ae-b0f9-a1bcc1b8cf14 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.457268] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.448s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.457500] env[61907]: DEBUG nova.objects.instance [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lazy-loading 'resources' on Instance uuid 23f9f3d1-b376-4502-927c-e9bd89693131 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.471866] env[61907]: INFO nova.scheduler.client.report [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Deleted allocations for instance cd2c70a7-a290-4d1a-91fe-5da772860603 [ 946.595488] env[61907]: DEBUG oslo_concurrency.lockutils [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.595798] env[61907]: DEBUG nova.compute.manager [req-f9f6c184-d00f-481c-8e0b-c5f022df1a24 req-d26c5c1b-cf44-4444-95c5-198c3a0268ba service nova] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Received event network-vif-deleted-5d7e9191-f828-48c6-81f9-0f30c6620477 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 946.639837] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52819399-5d4f-9008-b391-195d66ec3a49, 'name': SearchDatastore_Task, 'duration_secs': 0.03128} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.640650] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82de7c9d-f6ff-4953-bad0-a318bd6f7144 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.645717] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 946.645717] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524fcc8a-8620-4a6b-3ef0-be1fddab78c1" [ 946.645717] env[61907]: _type = "Task" [ 946.645717] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.653666] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524fcc8a-8620-4a6b-3ef0-be1fddab78c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.690323] env[61907]: DEBUG oslo_vmware.api [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Task: {'id': task-1244170, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.193235} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.690597] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.690801] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.690993] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.691196] env[61907]: INFO nova.compute.manager [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Took 1.06 seconds to destroy the instance on the hypervisor. [ 946.691452] env[61907]: DEBUG oslo.service.loopingcall [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.691654] env[61907]: DEBUG nova.compute.manager [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 946.691752] env[61907]: DEBUG nova.network.neutron [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 946.721492] env[61907]: DEBUG nova.network.neutron [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.957219] env[61907]: DEBUG nova.objects.instance [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.973066] env[61907]: INFO nova.compute.manager [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Took 17.93 seconds to build instance. [ 946.982504] env[61907]: DEBUG oslo_concurrency.lockutils [None req-581db555-d73a-4ef8-81ac-3b20cd7d9bfb tempest-ImagesOneServerNegativeTestJSON-1088355316 tempest-ImagesOneServerNegativeTestJSON-1088355316-project-member] Lock "cd2c70a7-a290-4d1a-91fe-5da772860603" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.880s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.160017] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524fcc8a-8620-4a6b-3ef0-be1fddab78c1, 'name': SearchDatastore_Task, 'duration_secs': 0.008861} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.160465] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.160716] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f50bf687-609c-44e1-bd75-cf4efaeadb71/f50bf687-609c-44e1-bd75-cf4efaeadb71.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 947.161055] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d37d7f41-1fdf-45c2-b120-dabb762ee111 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.168740] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 947.168740] env[61907]: value = "task-1244171" [ 947.168740] env[61907]: _type = "Task" [ 947.168740] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.179334] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244171, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.208312] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4774d411-3243-43ab-bf16-bec9ff370a63 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.217182] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36eebe28-297c-4a81-984f-6c31d19b39af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.223666] env[61907]: DEBUG nova.network.neutron [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.253912] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66669a0a-9624-48b5-955f-9ee3bd54a4e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.262265] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fd482f-adb3-4349-87a5-e110f2edb24f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.277424] env[61907]: DEBUG nova.compute.provider_tree [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.476490] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7074a14a-ed28-4244-9a31-0a9ad82e9a22 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.438s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.680310] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244171, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.448411} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.680600] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f50bf687-609c-44e1-bd75-cf4efaeadb71/f50bf687-609c-44e1-bd75-cf4efaeadb71.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 947.680865] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.681160] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c93e46b-04fc-4d1e-a535-3daf70b88c45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.688601] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 947.688601] env[61907]: value = "task-1244172" [ 947.688601] env[61907]: _type = "Task" [ 947.688601] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.692609] env[61907]: INFO nova.compute.manager [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Unrescuing [ 947.692987] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.693940] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.694091] env[61907]: DEBUG nova.network.neutron [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 947.701481] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244172, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.725545] env[61907]: INFO nova.compute.manager [-] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Took 1.03 seconds to deallocate network for instance. [ 947.801449] env[61907]: ERROR nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [req-d369330c-8112-4ac9-9e8d-27bb30780e96] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-d369330c-8112-4ac9-9e8d-27bb30780e96"}]} [ 947.825206] env[61907]: DEBUG nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 947.843545] env[61907]: DEBUG nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 947.843896] env[61907]: DEBUG nova.compute.provider_tree [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 947.862899] env[61907]: DEBUG nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 947.865249] env[61907]: DEBUG oslo_concurrency.lockutils [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.882498] env[61907]: DEBUG nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 947.973026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e8b1164-3880-4f94-9b59-cfbfcb506f2f tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.300s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.973026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.106s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.973026] env[61907]: DEBUG nova.compute.manager [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 947.973026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76581efb-2979-43d4-954b-1bfebb544e83 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.981893] env[61907]: DEBUG nova.compute.manager [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 947.983235] env[61907]: DEBUG nova.objects.instance [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.111597] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bc8052a-7a4e-427f-8b3a-3f44d5c6a959 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.125422] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208b7976-a94f-444e-85f8-0ed75e6fc927 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.161048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abe1d66-bb57-407f-a208-8e419742882b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.169985] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-362bde46-3646-4b8a-addc-70efcd63204b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.184400] env[61907]: DEBUG nova.compute.provider_tree [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.200190] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244172, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063304} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.200741] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.201887] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78221459-d4f1-4c63-bb1c-5667e5179cd9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.226382] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] f50bf687-609c-44e1-bd75-cf4efaeadb71/f50bf687-609c-44e1-bd75-cf4efaeadb71.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.227357] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca79882b-8151-45f4-b050-1ce5540b2448 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.243491] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.249555] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 948.249555] env[61907]: value = "task-1244173" [ 948.249555] env[61907]: _type = "Task" [ 948.249555] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.267463] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244173, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.687676] env[61907]: DEBUG nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 948.763850] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244173, 'name': ReconfigVM_Task, 'duration_secs': 0.315362} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.764605] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfigured VM instance instance-00000058 to attach disk [datastore1] f50bf687-609c-44e1-bd75-cf4efaeadb71/f50bf687-609c-44e1-bd75-cf4efaeadb71.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.764823] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b467bcf2-513f-4052-bbd8-654ab8623a3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.772611] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 948.772611] env[61907]: value = "task-1244174" [ 948.772611] env[61907]: _type = "Task" [ 948.772611] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.785237] env[61907]: DEBUG nova.network.neutron [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.791800] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244174, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.993699] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 948.994457] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-715af221-411c-4eb8-9e31-63b13f5580ca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.003356] env[61907]: DEBUG oslo_vmware.api [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 949.003356] env[61907]: value = "task-1244175" [ 949.003356] env[61907]: _type = "Task" [ 949.003356] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.011444] env[61907]: DEBUG oslo_vmware.api [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244175, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.164782] env[61907]: DEBUG nova.compute.manager [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 949.195589] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.738s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.198770] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.516s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.198770] env[61907]: DEBUG nova.objects.instance [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lazy-loading 'resources' on Instance uuid a31110ed-c4bb-4492-95a7-0a0ef5010c28 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.221530] env[61907]: INFO nova.scheduler.client.report [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted allocations for instance 23f9f3d1-b376-4502-927c-e9bd89693131 [ 949.283031] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244174, 'name': Rename_Task, 'duration_secs': 0.152905} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.283435] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 949.283757] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4706ff55-a1e0-4647-adfc-43fba26b212d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.290370] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 949.290370] env[61907]: value = "task-1244176" [ 949.290370] env[61907]: _type = "Task" [ 949.290370] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.296443] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.297162] env[61907]: DEBUG nova.objects.instance [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'flavor' on Instance uuid 53176c9e-d15c-49d5-b4a9-22b780279ecb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.304254] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244176, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.520025] env[61907]: DEBUG oslo_vmware.api [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244175, 'name': PowerOffVM_Task, 'duration_secs': 0.36714} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.520025] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 949.520025] env[61907]: DEBUG nova.compute.manager [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 949.520025] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c344c07-70a0-4d0c-9d82-fc2c4600c798 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.690075] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.730338] env[61907]: DEBUG oslo_concurrency.lockutils [None req-84778b7a-66e6-49ee-9014-7e044a89bf50 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "23f9f3d1-b376-4502-927c-e9bd89693131" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.165s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.805206] env[61907]: DEBUG oslo_vmware.api [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244176, 'name': PowerOnVM_Task, 'duration_secs': 0.464614} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.805822] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 949.806061] env[61907]: INFO nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Took 6.86 seconds to spawn the instance on the hypervisor. [ 949.806244] env[61907]: DEBUG nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 949.807032] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aabd84f-1e75-42a4-b524-1334831e69a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.813333] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc6f856-fc9c-4c70-b8b0-6710912c84d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.838486] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 949.843388] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-589510a2-19fb-4fc8-b7b1-b537fa9dcbc2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.853171] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 949.853171] env[61907]: value = "task-1244177" [ 949.853171] env[61907]: _type = "Task" [ 949.853171] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.862394] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.947409] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa137156-3389-48f5-8b2c-22bc8da2f29f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.956168] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bfe494-024c-441c-994a-f20edbabf0db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.991729] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d676b6-0df8-4ffe-97c0-486ff09f480b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.999685] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4ab452-549c-4fa2-9eed-1c4764ad5946 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.015899] env[61907]: DEBUG nova.compute.provider_tree [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.032486] env[61907]: DEBUG oslo_concurrency.lockutils [None req-278e0393-7e64-4f21-9bc3-b00e19422de1 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.061s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.373033] env[61907]: INFO nova.compute.manager [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Took 18.06 seconds to build instance. [ 950.384093] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244177, 'name': PowerOffVM_Task, 'duration_secs': 0.413834} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.384093] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 950.390515] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfiguring VM instance instance-0000004e to detach disk 2002 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 950.391502] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4845b9a3-a876-4754-a46f-189d30cedd30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.415820] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 950.415820] env[61907]: value = "task-1244178" [ 950.415820] env[61907]: _type = "Task" [ 950.415820] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.425814] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244178, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.519424] env[61907]: DEBUG nova.scheduler.client.report [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 950.593124] env[61907]: DEBUG nova.objects.instance [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 950.854512] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "e52648ac-e068-4a63-baa0-a1c34df52197" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.854929] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.855171] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.855363] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.855543] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.858214] env[61907]: INFO nova.compute.manager [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Terminating instance [ 950.873568] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a7479c7e-9985-47e0-945c-92810262dc11 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.575s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.930775] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244178, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.025700] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.028390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.079s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.052101] env[61907]: INFO nova.scheduler.client.report [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Deleted allocations for instance a31110ed-c4bb-4492-95a7-0a0ef5010c28 [ 951.099135] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.099343] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.099531] env[61907]: DEBUG nova.network.neutron [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.099779] env[61907]: DEBUG nova.objects.instance [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'info_cache' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.361904] env[61907]: DEBUG nova.compute.manager [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 951.362286] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 951.363135] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ed7f285-2457-4a41-be33-bd23d3d3cbd6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.372214] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 951.372959] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fad18489-731d-46d4-bf5c-f6af1bea2a4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.379468] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 951.379468] env[61907]: value = "task-1244179" [ 951.379468] env[61907]: _type = "Task" [ 951.379468] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.388745] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244179, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.428659] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244178, 'name': ReconfigVM_Task, 'duration_secs': 0.579649} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.428965] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfigured VM instance instance-0000004e to detach disk 2002 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 951.430946] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 951.431312] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ef11246-f637-4b7b-a8ab-874b5d0d55e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.438498] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 951.438498] env[61907]: value = "task-1244180" [ 951.438498] env[61907]: _type = "Task" [ 951.438498] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.448305] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244180, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.534671] env[61907]: INFO nova.compute.claims [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.549442] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "6a092752-058f-4196-848e-58a94d084cc7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.549840] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.569397] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6484b170-a33b-48a5-84f4-d652853cffb9 tempest-ServersTestJSON-1432564966 tempest-ServersTestJSON-1432564966-project-member] Lock "a31110ed-c4bb-4492-95a7-0a0ef5010c28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.600s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.604138] env[61907]: DEBUG nova.objects.base [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Object Instance<4c70ef23-fd26-4e90-b227-13aa050ff46d> lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 951.894936] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244179, 'name': PowerOffVM_Task, 'duration_secs': 0.320161} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.895322] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 951.895536] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 951.895859] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d9cfc068-aa34-433e-a2cc-13cd9a5edbe8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.953020] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244180, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.983212] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 951.983212] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 951.983212] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleting the datastore file [datastore2] e52648ac-e068-4a63-baa0-a1c34df52197 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.983212] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e81dc030-3ee6-41b8-a280-e541044820db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.992427] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 951.992427] env[61907]: value = "task-1244182" [ 951.992427] env[61907]: _type = "Task" [ 951.992427] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.001265] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244182, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.040817] env[61907]: INFO nova.compute.resource_tracker [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating resource usage from migration 24370906-b77f-4fe5-9015-61f8ba73b556 [ 952.053099] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 952.319391] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86768de-da10-410b-b34d-4a84baa22b59 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.329784] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91482bc3-3731-4c62-808d-7baba66bc063 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.365666] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c9f235-279a-42d9-b539-0ee2fd80cb59 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.370816] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa6bd41-eec4-436c-9169-38d35f005c3e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.385033] env[61907]: DEBUG nova.compute.provider_tree [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.451955] env[61907]: DEBUG oslo_vmware.api [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244180, 'name': PowerOnVM_Task, 'duration_secs': 0.605439} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.451955] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 952.452139] env[61907]: DEBUG nova.compute.manager [None req-3a89822d-9e76-4c47-bda0-ab074d98eca9 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 952.453026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6d437d-6f5b-4c49-932b-d97a878dc33d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.504570] env[61907]: DEBUG oslo_vmware.api [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244182, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414309} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.504570] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 952.504570] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 952.504570] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 952.504967] env[61907]: INFO nova.compute.manager [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Took 1.14 seconds to destroy the instance on the hypervisor. [ 952.505206] env[61907]: DEBUG oslo.service.loopingcall [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.505398] env[61907]: DEBUG nova.compute.manager [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 952.505538] env[61907]: DEBUG nova.network.neutron [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 952.583993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.675741] env[61907]: DEBUG nova.compute.manager [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 952.675741] env[61907]: DEBUG nova.compute.manager [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 952.675741] env[61907]: DEBUG oslo_concurrency.lockutils [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.675741] env[61907]: DEBUG oslo_concurrency.lockutils [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.675741] env[61907]: DEBUG nova.network.neutron [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.888131] env[61907]: DEBUG nova.scheduler.client.report [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 953.396066] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.368s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.396439] env[61907]: INFO nova.compute.manager [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Migrating [ 953.409754] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.167s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.410316] env[61907]: DEBUG nova.objects.instance [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lazy-loading 'resources' on Instance uuid c3bf9001-7166-4064-89d5-96be073fa7c2 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.442577] env[61907]: DEBUG nova.network.neutron [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [{"id": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "address": "fa:16:3e:2c:b6:15", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5290ad3b-91", "ovs_interfaceid": "5290ad3b-915d-49f5-b5e8-57438dd6ca7b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.460459] env[61907]: DEBUG nova.network.neutron [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.460459] env[61907]: DEBUG nova.network.neutron [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.922252] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.922252] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.922252] env[61907]: DEBUG nova.network.neutron [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.945742] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "refresh_cache-4c70ef23-fd26-4e90-b227-13aa050ff46d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.963213] env[61907]: DEBUG oslo_concurrency.lockutils [req-bfe2d06b-3ce7-42e9-8f4a-73b1dc909e1c req-2d5d4465-defc-460a-bd22-40621d73b81d service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.010508] env[61907]: DEBUG nova.network.neutron [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.103239] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.103822] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.197735] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04622e7b-3e0e-4220-87da-09638202a9b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.207074] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1e3c25-5141-4727-8e36-afaf3d1429e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.245501] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb0ce6d-1dea-42e4-bfbf-13ebcdb31beb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.254485] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba14a1b-82ba-4b00-9054-85e3c99bbea4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.270433] env[61907]: DEBUG nova.compute.provider_tree [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.515528] env[61907]: INFO nova.compute.manager [-] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Took 2.01 seconds to deallocate network for instance. [ 954.607957] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 954.724141] env[61907]: DEBUG nova.compute.manager [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 954.724914] env[61907]: DEBUG nova.compute.manager [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing instance network info cache due to event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 954.724914] env[61907]: DEBUG oslo_concurrency.lockutils [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.724914] env[61907]: DEBUG oslo_concurrency.lockutils [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.725112] env[61907]: DEBUG nova.network.neutron [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 954.800662] env[61907]: ERROR nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] [req-84a332c6-9311-4d9f-a43e-8fa9b2f2acac] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-84a332c6-9311-4d9f-a43e-8fa9b2f2acac"}]} [ 954.820474] env[61907]: DEBUG nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 954.826208] env[61907]: DEBUG nova.network.neutron [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.841418] env[61907]: DEBUG nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 954.841659] env[61907]: DEBUG nova.compute.provider_tree [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 954.855905] env[61907]: DEBUG nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 954.878995] env[61907]: DEBUG nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 954.952215] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 954.956134] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1c8dfea-7f22-4715-bf04-473492033e66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.965056] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 954.965056] env[61907]: value = "task-1244183" [ 954.965056] env[61907]: _type = "Task" [ 954.965056] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.975397] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.023967] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.133677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.145907] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5084abc-7614-40c8-a964-cdd03e179889 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.153901] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3683ff69-8d18-4784-a08a-956bdc81a839 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.186087] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b83fc0-d9d6-4be8-b965-28cccc41eff3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.196976] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed8a511b-0cac-4404-9450-7d89451427b9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.216995] env[61907]: DEBUG nova.compute.provider_tree [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.329936] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.482020] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.527389] env[61907]: DEBUG nova.network.neutron [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updated VIF entry in instance network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 955.527949] env[61907]: DEBUG nova.network.neutron [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.619047] env[61907]: DEBUG nova.compute.manager [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 955.619242] env[61907]: DEBUG nova.compute.manager [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing instance network info cache due to event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 955.620228] env[61907]: DEBUG oslo_concurrency.lockutils [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.620333] env[61907]: DEBUG oslo_concurrency.lockutils [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.620489] env[61907]: DEBUG nova.network.neutron [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 955.762016] env[61907]: DEBUG nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 112 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 955.762016] env[61907]: DEBUG nova.compute.provider_tree [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 112 to 113 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 955.762016] env[61907]: DEBUG nova.compute.provider_tree [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 955.976517] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.035892] env[61907]: DEBUG oslo_concurrency.lockutils [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.036250] env[61907]: DEBUG nova.compute.manager [req-715bb9f8-7ff4-4c47-bd83-470751ae180f req-f3a815dc-50e5-4f95-bfa4-b34463e0770f service nova] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Received event network-vif-deleted-9897ff2a-3de3-42a2-af12-5f0ef6234904 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 956.266388] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.857s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.271470] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.579s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.293522] env[61907]: INFO nova.scheduler.client.report [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Deleted allocations for instance c3bf9001-7166-4064-89d5-96be073fa7c2 [ 956.425848] env[61907]: DEBUG nova.network.neutron [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updated VIF entry in instance network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 956.425848] env[61907]: DEBUG nova.network.neutron [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.477203] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.687431] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.687678] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.776156] env[61907]: INFO nova.compute.claims [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 956.804799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a9364055-8d6b-4411-9acf-30d65807509f tempest-ServerShowV247Test-2000203089 tempest-ServerShowV247Test-2000203089-project-member] Lock "c3bf9001-7166-4064-89d5-96be073fa7c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.786s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.847900] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9b20d4-a53f-481e-9987-2ec52c5f5bb4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.868158] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 956.928309] env[61907]: DEBUG oslo_concurrency.lockutils [req-384ee743-f23b-4ecc-b2ba-26af28634b63 req-199610c6-0f03-4b13-9f8f-69ce3832d247 service nova] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.977462] env[61907]: DEBUG oslo_vmware.api [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244183, 'name': PowerOnVM_Task, 'duration_secs': 1.681349} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.977749] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 956.977952] env[61907]: DEBUG nova.compute.manager [None req-c0ee5e80-a66c-48cc-b03b-114cbf0c4b00 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 956.978887] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33983037-33c3-4fb1-b77d-f394ab43ba65 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.189681] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 957.281948] env[61907]: INFO nova.compute.resource_tracker [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating resource usage from migration 10a7713a-f8c3-42d0-b6ef-9f707de1f035 [ 957.377154] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 957.377154] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17d8f38b-728e-4bf7-bbaf-a0c0fa5de48c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.387236] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 957.387236] env[61907]: value = "task-1244184" [ 957.387236] env[61907]: _type = "Task" [ 957.387236] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.397584] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.503792] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdaebbf-26db-4afe-aa7c-e52abe87e601 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.512662] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f684699-5fcd-4504-a124-4e55cd00e60e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.544141] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677fc051-66b6-4740-b934-37fc3df280b4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.551595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebffeb82-ec44-459e-976f-510dd893402b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.565331] env[61907]: DEBUG nova.compute.provider_tree [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.651026] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 957.651591] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing instance network info cache due to event network-changed-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 957.651834] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquiring lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.651981] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquired lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.652157] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Refreshing network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 957.709186] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.898052] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244184, 'name': PowerOffVM_Task, 'duration_secs': 0.227027} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.898052] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.898052] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 958.068278] env[61907]: DEBUG nova.scheduler.client.report [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 958.404870] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.405259] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.405492] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.405741] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.406032] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.406250] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.406471] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.406641] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.406814] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.406983] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.407181] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.412607] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75fb3fcf-ce47-4951-a61b-e5e7089f77f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.430826] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 958.430826] env[61907]: value = "task-1244185" [ 958.430826] env[61907]: _type = "Task" [ 958.430826] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.439347] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244185, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.573390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.305s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.573603] env[61907]: INFO nova.compute.manager [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Migrating [ 958.580292] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.997s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.581732] env[61907]: INFO nova.compute.claims [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.640802] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updated VIF entry in instance network info cache for port 8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.641196] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [{"id": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "address": "fa:16:3e:4c:b4:63", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.193", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8c5ce2c9-4b", "ovs_interfaceid": "8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.941372] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244185, 'name': ReconfigVM_Task, 'duration_secs': 0.248516} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.941733] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 959.094065] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.094300] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.094499] env[61907]: DEBUG nova.network.neutron [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 959.143762] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Releasing lock "refresh_cache-53176c9e-d15c-49d5-b4a9-22b780279ecb" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.143762] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 959.143992] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing instance network info cache due to event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 959.144265] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.144449] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.144657] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.448458] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.448758] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.448901] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.449196] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.449393] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.449586] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.449832] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.450063] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.450286] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.450493] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.450710] env[61907]: DEBUG nova.virt.hardware [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.456132] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 959.456462] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6d397e4f-7c7e-4c31-b846-f47e8759934b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.476156] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 959.476156] env[61907]: value = "task-1244186" [ 959.476156] env[61907]: _type = "Task" [ 959.476156] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.484476] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244186, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.697048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "4430db3b-0276-42b6-a67e-386a164ddc0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.697048] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.825941] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81dc483f-48ac-4da6-a1f4-258281dd9639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.834139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57c5dbcb-9a37-400d-b7ed-0044f61cccc1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.866843] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22f3546-1ef3-4758-be34-5fe1b6cb4146 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.874124] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67ae1e9a-4fd4-42e4-bc73-a5b2166e0c6f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.878536] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updated VIF entry in instance network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.878906] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.880583] env[61907]: DEBUG nova.network.neutron [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.891507] env[61907]: DEBUG nova.compute.provider_tree [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.985807] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244186, 'name': ReconfigVM_Task, 'duration_secs': 0.208672} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.986173] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 959.986860] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a8b7e3-1035-4d2f-9d27-2f3671e674be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.009196] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 960.009429] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1926349a-a622-4534-b977-88c5f75104bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.025791] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 960.025791] env[61907]: value = "task-1244187" [ 960.025791] env[61907]: _type = "Task" [ 960.025791] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.033109] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.199018] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 960.383133] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.383448] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 960.383649] env[61907]: DEBUG nova.compute.manager [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 960.384104] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.384288] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.384698] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 960.386487] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.394495] env[61907]: DEBUG nova.scheduler.client.report [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 960.535486] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.752500] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.899227] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.899764] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 960.902105] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.878s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.902331] env[61907]: DEBUG nova.objects.instance [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lazy-loading 'resources' on Instance uuid e52648ac-e068-4a63-baa0-a1c34df52197 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.037178] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244187, 'name': ReconfigVM_Task, 'duration_secs': 0.75667} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.037546] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Reconfigured VM instance instance-00000056 to attach disk [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f/f87e45da-187b-4aad-b7bf-b4228a8b2f1f.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 961.037757] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.153706] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 961.154094] env[61907]: DEBUG nova.network.neutron [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.405642] env[61907]: DEBUG nova.compute.utils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 961.409731] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 961.409899] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 961.450297] env[61907]: DEBUG nova.policy [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8682003dee14f25ba212739afd3d335', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6a88771a8f84196ae271dbd57590c15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 961.546053] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49fed482-f24c-45a7-8789-55d0241bd69b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.571849] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d515605-5cb7-4e18-9d5c-16c8829bdc25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.590485] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 961.623611] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de935590-1aeb-4af2-b079-e5d509bf1f9a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.631112] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee67817c-859e-402d-abaa-962326f19294 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.662017] env[61907]: DEBUG oslo_concurrency.lockutils [req-1d523263-46f7-486a-8232-065faa21984f req-5125dea1-67e4-42bd-9006-bd7af5c3cd6e service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.663151] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48686768-2536-4c13-99dc-a425b6646ce0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.672663] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e5d461-7b09-461a-a784-f4308ba00c46 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.689456] env[61907]: DEBUG nova.compute.provider_tree [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.717739] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Successfully created port: 88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.904106] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c273b9f0-c860-4d88-ada2-9aedca6e3e30 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.910065] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 961.929609] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.127607] env[61907]: DEBUG nova.network.neutron [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Port 8b42cc64-101e-4eec-953d-5c2cc55dd252 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 962.191098] env[61907]: DEBUG nova.scheduler.client.report [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 962.439976] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.440526] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-50ca19c5-3369-400d-b35a-1fc616445113 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.448346] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 962.448346] env[61907]: value = "task-1244188" [ 962.448346] env[61907]: _type = "Task" [ 962.448346] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.458857] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244188, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.695425] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.793s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.697715] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.564s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.699202] env[61907]: INFO nova.compute.claims [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.716482] env[61907]: INFO nova.scheduler.client.report [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted allocations for instance e52648ac-e068-4a63-baa0-a1c34df52197 [ 962.942068] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 962.958236] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244188, 'name': PowerOffVM_Task, 'duration_secs': 0.327405} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.958514] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 962.958711] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 962.969300] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.969549] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.969713] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.969903] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.970069] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.970223] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.970500] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.970684] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.970856] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.971034] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.971216] env[61907]: DEBUG nova.virt.hardware [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.972014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca65b77-f17a-4721-8deb-1769711d7668 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.979382] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfb44fb-e90e-4421-9eb9-0354eb0b4f0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.149874] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.150221] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.150221] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.224081] env[61907]: DEBUG nova.compute.manager [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Received event network-vif-plugged-88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 963.224343] env[61907]: DEBUG oslo_concurrency.lockutils [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] Acquiring lock "6a092752-058f-4196-848e-58a94d084cc7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.224558] env[61907]: DEBUG oslo_concurrency.lockutils [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] Lock "6a092752-058f-4196-848e-58a94d084cc7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.224738] env[61907]: DEBUG oslo_concurrency.lockutils [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] Lock "6a092752-058f-4196-848e-58a94d084cc7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.224944] env[61907]: DEBUG nova.compute.manager [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] No waiting events found dispatching network-vif-plugged-88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 963.225162] env[61907]: WARNING nova.compute.manager [req-7465994c-fed1-4fb5-bc81-2dc4f716aeb3 req-8921e599-1b1f-4d60-8cae-ccd3da4050c8 service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Received unexpected event network-vif-plugged-88b21882-cb3c-444d-8966-d16967161a8a for instance with vm_state building and task_state spawning. [ 963.227806] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f331da6a-bbea-4ab4-b0af-857245914130 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "e52648ac-e068-4a63-baa0-a1c34df52197" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.373s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.305751] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Successfully updated port: 88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 963.465889] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.466173] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.466341] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.466530] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.466679] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.466828] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.467047] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.467222] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.467423] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.467556] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.467733] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.472673] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6700eb3d-6d9a-4726-a060-7da2252ac8d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.492763] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 963.492763] env[61907]: value = "task-1244189" [ 963.492763] env[61907]: _type = "Task" [ 963.492763] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.501388] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244189, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.808581] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.808778] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.808873] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 963.887745] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe996f4-50f7-4cc5-94f9-da1c63a6d49c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.895553] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c3ae08-d2dd-4134-8b4a-566cce25238c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.924715] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad74e216-96e2-463c-869f-8c6a9ee6cbb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.931985] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7840212-859a-4234-a9be-06d7f310ca75 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.947101] env[61907]: DEBUG nova.compute.provider_tree [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.002798] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244189, 'name': ReconfigVM_Task, 'duration_secs': 0.258062} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.003135] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 964.185026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.185290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.185451] env[61907]: DEBUG nova.network.neutron [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 964.341908] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 964.450610] env[61907]: DEBUG nova.scheduler.client.report [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 964.467333] env[61907]: DEBUG nova.network.neutron [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Updating instance_info_cache with network_info: [{"id": "88b21882-cb3c-444d-8966-d16967161a8a", "address": "fa:16:3e:20:38:e4", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b21882-cb", "ovs_interfaceid": "88b21882-cb3c-444d-8966-d16967161a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.510596] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.510920] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.511224] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.511495] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.511719] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.511946] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.512246] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.512484] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.512738] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.513085] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.513356] env[61907]: DEBUG nova.virt.hardware [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.520060] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfiguring VM instance instance-00000057 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 964.520615] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-286fcf6e-b4f8-4161-a525-88caabb9356a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.539776] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 964.539776] env[61907]: value = "task-1244190" [ 964.539776] env[61907]: _type = "Task" [ 964.539776] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.548263] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244190, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.873145] env[61907]: DEBUG nova.network.neutron [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.955310] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.955829] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 964.958246] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.249s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.959743] env[61907]: INFO nova.compute.claims [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.969509] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.969789] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Instance network_info: |[{"id": "88b21882-cb3c-444d-8966-d16967161a8a", "address": "fa:16:3e:20:38:e4", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b21882-cb", "ovs_interfaceid": "88b21882-cb3c-444d-8966-d16967161a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 964.970173] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:20:38:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '503991c4-44d0-42d9-aa03-5259331f1051', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88b21882-cb3c-444d-8966-d16967161a8a', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.977700] env[61907]: DEBUG oslo.service.loopingcall [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.977970] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 964.978317] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-edba2d02-c70e-4041-8b1d-27b1bcc21d0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.999699] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.999699] env[61907]: value = "task-1244191" [ 964.999699] env[61907]: _type = "Task" [ 964.999699] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.008068] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244191, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.048843] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244190, 'name': ReconfigVM_Task, 'duration_secs': 0.153572} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.049161] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfigured VM instance instance-00000057 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 965.049953] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dde2c17-248b-43a2-8b0d-13579a42305f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.071856] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.072567] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-205802a0-08c1-4ee6-a201-1f19f0f2b445 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.092892] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 965.092892] env[61907]: value = "task-1244192" [ 965.092892] env[61907]: _type = "Task" [ 965.092892] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.103850] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244192, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.250021] env[61907]: DEBUG nova.compute.manager [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Received event network-changed-88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 965.250293] env[61907]: DEBUG nova.compute.manager [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Refreshing instance network info cache due to event network-changed-88b21882-cb3c-444d-8966-d16967161a8a. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 965.250483] env[61907]: DEBUG oslo_concurrency.lockutils [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] Acquiring lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.250633] env[61907]: DEBUG oslo_concurrency.lockutils [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] Acquired lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.250795] env[61907]: DEBUG nova.network.neutron [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Refreshing network info cache for port 88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 965.376098] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.464320] env[61907]: DEBUG nova.compute.utils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.468177] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 965.468341] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 965.510664] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244191, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.527470] env[61907]: DEBUG nova.policy [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '457576b3cac24f2bb81eee33c2b52aae', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a2fb5c5950d5400aa93060c09137c247', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.602638] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244192, 'name': ReconfigVM_Task, 'duration_secs': 0.255346} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.602941] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d/5475c612-e718-49eb-9760-9cfedbd7931d.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.603312] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.688829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.689127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.689521] env[61907]: DEBUG nova.objects.instance [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'flavor' on Instance uuid fcde07bf-858a-4377-b27c-4f17356306af {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.876612] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Successfully created port: b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.911853] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cba849d-4fe4-4278-9fce-8b153f84d577 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.933439] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ed013a-c125-4f89-aa02-93b166a937d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.940430] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 965.968735] env[61907]: DEBUG nova.network.neutron [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Updated VIF entry in instance network info cache for port 88b21882-cb3c-444d-8966-d16967161a8a. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 965.969108] env[61907]: DEBUG nova.network.neutron [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Updating instance_info_cache with network_info: [{"id": "88b21882-cb3c-444d-8966-d16967161a8a", "address": "fa:16:3e:20:38:e4", "network": {"id": "3348260f-9176-4421-be00-0f535d09cb03", "bridge": "br-int", "label": "tempest-ImagesTestJSON-536877245-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6a88771a8f84196ae271dbd57590c15", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "503991c4-44d0-42d9-aa03-5259331f1051", "external-id": "nsx-vlan-transportzone-3", "segmentation_id": 3, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88b21882-cb", "ovs_interfaceid": "88b21882-cb3c-444d-8966-d16967161a8a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.970463] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 966.021031] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244191, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.109872] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b440b4c8-274d-41d7-8ea4-29d51db331d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.133181] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-398febef-9f1e-4b15-9c83-21c13d9bc803 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.150454] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.184250] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95bf1a5-2bdb-4d45-8a5a-515e50571cb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.192015] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf31befd-fbf7-4823-90e0-96077e5d728a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.227033] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34510ca-10c1-4b10-acc1-0edde9d4b9fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.235504] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574c8fe2-eb91-45cb-9200-1fcaadda0735 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.250415] env[61907]: DEBUG nova.compute.provider_tree [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.298196] env[61907]: DEBUG nova.objects.instance [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'pci_requests' on Instance uuid fcde07bf-858a-4377-b27c-4f17356306af {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.447404] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 966.447721] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fa62622c-d748-4e60-8051-df3024305965 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.454393] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 966.454393] env[61907]: value = "task-1244193" [ 966.454393] env[61907]: _type = "Task" [ 966.454393] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.462222] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244193, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.473897] env[61907]: DEBUG oslo_concurrency.lockutils [req-443d1392-d882-4fee-a2d0-658e02a65974 req-f8dbee0d-c443-4d88-9905-621cfb85aa6f service nova] Releasing lock "refresh_cache-6a092752-058f-4196-848e-58a94d084cc7" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.510937] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244191, 'name': CreateVM_Task, 'duration_secs': 1.304076} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.511783] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 966.512519] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.512699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.513078] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.513612] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-762b376c-0603-40cc-bc5f-17a152e00d3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.518322] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 966.518322] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528297b1-f705-ad0e-909b-39205eb8271a" [ 966.518322] env[61907]: _type = "Task" [ 966.518322] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.525957] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528297b1-f705-ad0e-909b-39205eb8271a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.692290] env[61907]: DEBUG nova.network.neutron [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Port e44f586e-da28-4e87-95ea-518bbf489218 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 966.753175] env[61907]: DEBUG nova.scheduler.client.report [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 966.801093] env[61907]: DEBUG nova.objects.base [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 966.801374] env[61907]: DEBUG nova.network.neutron [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.862882] env[61907]: DEBUG nova.policy [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.966385] env[61907]: DEBUG oslo_vmware.api [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244193, 'name': PowerOnVM_Task, 'duration_secs': 0.359206} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.966699] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 966.966893] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2a1a8fed-650e-40c9-8666-55ec58310cbb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance 'f87e45da-187b-4aad-b7bf-b4228a8b2f1f' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 966.981720] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.008967] env[61907]: DEBUG nova.virt.hardware [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.009688] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d7121e0-c9e4-4230-93eb-fdbfb440f12c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.019744] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfa9b7f1-2d73-4fcb-bb05-d70f537d7dde {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.039852] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528297b1-f705-ad0e-909b-39205eb8271a, 'name': SearchDatastore_Task, 'duration_secs': 0.009665} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.040224] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.040519] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.040838] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.040968] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.041814] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.041814] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5fb466ca-b369-4999-9033-fbda7b985581 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.049793] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.050028] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 967.050751] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fde4293d-330d-4b07-8cc4-6c5cc69d5db7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.055790] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 967.055790] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d67be-9d2a-64ab-f7a5-985a565fd709" [ 967.055790] env[61907]: _type = "Task" [ 967.055790] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.063204] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d67be-9d2a-64ab-f7a5-985a565fd709, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.257820] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.258443] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 967.261081] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.509s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.262522] env[61907]: INFO nova.compute.claims [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 967.274452] env[61907]: DEBUG nova.compute.manager [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-vif-plugged-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 967.274693] env[61907]: DEBUG oslo_concurrency.lockutils [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] Acquiring lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.274970] env[61907]: DEBUG oslo_concurrency.lockutils [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.275162] env[61907]: DEBUG oslo_concurrency.lockutils [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.275362] env[61907]: DEBUG nova.compute.manager [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] No waiting events found dispatching network-vif-plugged-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 967.275534] env[61907]: WARNING nova.compute.manager [req-2424d0c7-68e8-4c39-a8e2-246fb2116058 req-652e17e3-3dcd-443c-be1a-a8b2f03c707e service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received unexpected event network-vif-plugged-b79b943f-5924-4ec0-bf11-eda9d8da3285 for instance with vm_state building and task_state spawning. [ 967.340347] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Successfully updated port: b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 967.566872] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d67be-9d2a-64ab-f7a5-985a565fd709, 'name': SearchDatastore_Task, 'duration_secs': 0.008194} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.567757] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3943a439-2514-4c82-9635-9941907ddd99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.572540] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 967.572540] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525e9646-73c1-9bc9-7851-ec5c9a5e140a" [ 967.572540] env[61907]: _type = "Task" [ 967.572540] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.579698] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525e9646-73c1-9bc9-7851-ec5c9a5e140a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.718118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.718526] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.718825] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.766643] env[61907]: DEBUG nova.compute.utils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.770277] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 967.770483] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 967.840261] env[61907]: DEBUG nova.policy [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b4762c5f79e48fb8a4f9701e5c59ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c30836c7053443aaab6e1dbf7783259', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.842099] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.842259] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.842370] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.082068] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525e9646-73c1-9bc9-7851-ec5c9a5e140a, 'name': SearchDatastore_Task, 'duration_secs': 0.009147} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.082351] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.082609] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6a092752-058f-4196-848e-58a94d084cc7/6a092752-058f-4196-848e-58a94d084cc7.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 968.082867] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f45a402-6ae9-4e16-97a8-14cb62a92837 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.088642] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 968.088642] env[61907]: value = "task-1244194" [ 968.088642] env[61907]: _type = "Task" [ 968.088642] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.096857] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244194, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.208904] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Successfully created port: ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 968.269465] env[61907]: DEBUG nova.compute.manager [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 968.269790] env[61907]: DEBUG oslo_concurrency.lockutils [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.270146] env[61907]: DEBUG oslo_concurrency.lockutils [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.270395] env[61907]: DEBUG oslo_concurrency.lockutils [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.270640] env[61907]: DEBUG nova.compute.manager [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] No waiting events found dispatching network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 968.270896] env[61907]: WARNING nova.compute.manager [req-022fce7e-8670-4bb5-9982-998a324a1d08 req-3e7b47f9-e16e-487c-a62d-ce54421b0b1b service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received unexpected event network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 for instance with vm_state active and task_state None. [ 968.276583] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 968.378922] env[61907]: DEBUG nova.network.neutron [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Successfully updated port: 404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.399643] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 968.539301] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c549904-75ba-4d86-a4d9-39a479303901 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.549455] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cc898d-f133-4e65-9659-e0764ab15c54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.586102] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec651441-e8b1-47ca-bab3-0919f8368bd4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.596139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84e0a47e-e51c-48a1-9834-314d8e1b5693 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.603219] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244194, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.613218] env[61907]: DEBUG nova.compute.provider_tree [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.627141] env[61907]: DEBUG nova.network.neutron [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.781113] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.781327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.781507] env[61907]: DEBUG nova.network.neutron [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.884808] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.884808] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.885124] env[61907]: DEBUG nova.network.neutron [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 969.099221] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244194, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521632} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.099531] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6a092752-058f-4196-848e-58a94d084cc7/6a092752-058f-4196-848e-58a94d084cc7.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 969.099751] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.100024] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-12b445d5-f225-4e29-aae7-d1d52e2f4b08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.108226] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 969.108226] env[61907]: value = "task-1244195" [ 969.108226] env[61907]: _type = "Task" [ 969.108226] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.118255] env[61907]: DEBUG nova.scheduler.client.report [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 969.122083] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244195, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.130138] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.130883] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Instance network_info: |[{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 969.131125] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9b:db:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e7a44713-0af1-486e-bc0d-00e03a769fa4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b79b943f-5924-4ec0-bf11-eda9d8da3285', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 969.138702] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Creating folder: Project (a2fb5c5950d5400aa93060c09137c247). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.139761] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6ccc44f0-2c67-4793-9afa-9e6ee33042b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.151478] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Created folder: Project (a2fb5c5950d5400aa93060c09137c247) in parent group-v268168. [ 969.151713] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Creating folder: Instances. Parent ref: group-v268298. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 969.152018] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f1a2e63-80fa-402e-aeb8-394bc2d4142e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.162671] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Created folder: Instances in parent group-v268298. [ 969.162958] env[61907]: DEBUG oslo.service.loopingcall [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.163173] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 969.163394] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41fe6afc-fe21-4261-85fd-1813ce97cdb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.183172] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 969.183172] env[61907]: value = "task-1244198" [ 969.183172] env[61907]: _type = "Task" [ 969.183172] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.191018] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244198, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.296755] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 969.329674] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 969.330149] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 969.330583] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 969.330918] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 969.331340] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 969.331625] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 969.331989] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 969.332943] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 969.332943] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 969.332943] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 969.333256] env[61907]: DEBUG nova.virt.hardware [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 969.335077] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393706be-ac82-4afc-96de-cff5dad59e4e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.350772] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1dfad2-4a75-407a-8c75-78779c4dbb21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.440373] env[61907]: WARNING nova.network.neutron [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] 9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b already exists in list: networks containing: ['9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b']. ignoring it [ 969.457597] env[61907]: DEBUG nova.compute.manager [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 969.457802] env[61907]: DEBUG nova.compute.manager [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing instance network info cache due to event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 969.457998] env[61907]: DEBUG oslo_concurrency.lockutils [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.458215] env[61907]: DEBUG oslo_concurrency.lockutils [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.458416] env[61907]: DEBUG nova.network.neutron [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.620114] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244195, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121212} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.623729] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.626256] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.365s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.627171] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 969.634935] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54470af6-6fce-466d-8e93-73085c6775df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.660753] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] 6a092752-058f-4196-848e-58a94d084cc7/6a092752-058f-4196-848e-58a94d084cc7.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.663985] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1439a576-f155-42d4-8583-c44d307fcdfd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.679218] env[61907]: DEBUG nova.network.neutron [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.692311] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 969.692311] env[61907]: value = "task-1244199" [ 969.692311] env[61907]: _type = "Task" [ 969.692311] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.698685] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244198, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.704295] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244199, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.799844] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.800148] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.800398] env[61907]: DEBUG nova.compute.manager [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Going to confirm migration 2 {{(pid=61907) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 969.809543] env[61907]: DEBUG nova.network.neutron [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "address": "fa:16:3e:4a:e6:f4", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404d6d57-b8", "ovs_interfaceid": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.881040] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Successfully updated port: ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 970.141139] env[61907]: DEBUG nova.compute.utils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 970.142559] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 970.142737] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 970.182566] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.191036] env[61907]: DEBUG nova.policy [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88b29a917e03445281f16fba6d536280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7652e98cde994af28b7bac0b81547474', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 970.203714] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244199, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.206832] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244198, 'name': CreateVM_Task, 'duration_secs': 0.666352} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.207646] env[61907]: DEBUG nova.network.neutron [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updated VIF entry in instance network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 970.207969] env[61907]: DEBUG nova.network.neutron [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.209754] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 970.209754] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.209888] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.210185] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 970.210652] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49505378-c897-4568-bb4f-c64dfe0c7248 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.215102] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 970.215102] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fe9765-e6d2-ddc5-2ca7-18f186b5dce8" [ 970.215102] env[61907]: _type = "Task" [ 970.215102] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.224220] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fe9765-e6d2-ddc5-2ca7-18f186b5dce8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.312125] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.312710] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.312871] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.313750] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb0a5c53-7599-4cbe-a5a7-ecc5fff4f8a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.333406] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 970.333681] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 970.333845] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 970.334359] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 970.334359] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 970.334359] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 970.334577] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 970.334717] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 970.334891] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 970.335087] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 970.335287] env[61907]: DEBUG nova.virt.hardware [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 970.341617] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfiguring VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 970.341951] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f30ffa42-4b35-4651-a78d-9c1847e4905b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.354627] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.354803] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.354985] env[61907]: DEBUG nova.network.neutron [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.355219] env[61907]: DEBUG nova.objects.instance [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'info_cache' on Instance uuid f87e45da-187b-4aad-b7bf-b4228a8b2f1f {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 970.362770] env[61907]: DEBUG oslo_vmware.api [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 970.362770] env[61907]: value = "task-1244200" [ 970.362770] env[61907]: _type = "Task" [ 970.362770] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.372026] env[61907]: DEBUG oslo_vmware.api [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244200, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.384281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.384281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.384399] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.397076] env[61907]: DEBUG nova.compute.manager [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 970.397156] env[61907]: DEBUG nova.compute.manager [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-404d6d57-b8d5-42c7-b255-d3b74e4f5797. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 970.397389] env[61907]: DEBUG oslo_concurrency.lockutils [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.397595] env[61907]: DEBUG oslo_concurrency.lockutils [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.397789] env[61907]: DEBUG nova.network.neutron [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port 404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.478127] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Successfully created port: bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 970.646496] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 970.706614] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244199, 'name': ReconfigVM_Task, 'duration_secs': 0.609207} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.706910] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Reconfigured VM instance instance-00000059 to attach disk [datastore2] 6a092752-058f-4196-848e-58a94d084cc7/6a092752-058f-4196-848e-58a94d084cc7.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.707559] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae4f3d2f-4b81-40af-b358-f9ea1339f885 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.712041] env[61907]: DEBUG oslo_concurrency.lockutils [req-75e7215c-a2ae-4ba9-b9d3-fa18f9ade7e7 req-b1be1769-ed36-47e7-8d71-791be7539842 service nova] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.715264] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 970.715264] env[61907]: value = "task-1244201" [ 970.715264] env[61907]: _type = "Task" [ 970.715264] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.718659] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a27164-3650-4445-a560-1850b4c50b9b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.730106] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fe9765-e6d2-ddc5-2ca7-18f186b5dce8, 'name': SearchDatastore_Task, 'duration_secs': 0.009939} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.746469] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.746772] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 970.747060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.747227] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.747438] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 970.747859] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244201, 'name': Rename_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.748170] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59d78690-a623-43e4-ae2a-073f88327173 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.750493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0ecedd-9c2a-4ebc-b071-f85be7ebb817 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.758324] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 970.763397] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 970.763584] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 970.764491] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0bf3a5df-8882-45cc-9236-15aad02ebd10 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.769624] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 970.769624] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce1d8-2cf0-20ba-d0eb-3814130eeb74" [ 970.769624] env[61907]: _type = "Task" [ 970.769624] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.777156] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce1d8-2cf0-20ba-d0eb-3814130eeb74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.874752] env[61907]: DEBUG oslo_vmware.api [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244200, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.928711] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 971.226589] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244201, 'name': Rename_Task, 'duration_secs': 0.129619} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.226589] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.226589] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-28ada3b9-7f24-46c0-99b6-4e30fa9890ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.233022] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 971.233022] env[61907]: value = "task-1244202" [ 971.233022] env[61907]: _type = "Task" [ 971.233022] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.238876] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244202, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.251576] env[61907]: DEBUG nova.network.neutron [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [{"id": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "address": "fa:16:3e:f2:c0:21", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped422e68-2f", "ovs_interfaceid": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.268029] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 971.268029] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b980b24-94c0-4e6e-916c-66045146b440 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.276312] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 971.276312] env[61907]: value = "task-1244203" [ 971.276312] env[61907]: _type = "Task" [ 971.276312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.287516] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244203, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.291122] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce1d8-2cf0-20ba-d0eb-3814130eeb74, 'name': SearchDatastore_Task, 'duration_secs': 0.008639} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.292039] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c3474a2-5d2b-4eef-845d-d591abf712aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.297975] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 971.297975] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528af506-c586-f45d-d4b2-c2427d29d00e" [ 971.297975] env[61907]: _type = "Task" [ 971.297975] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.306591] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528af506-c586-f45d-d4b2-c2427d29d00e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.373914] env[61907]: DEBUG oslo_vmware.api [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244200, 'name': ReconfigVM_Task, 'duration_secs': 0.756384} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.374994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.375457] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfigured VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 971.514537] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 971.566426] env[61907]: DEBUG nova.network.neutron [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port 404d6d57-b8d5-42c7-b255-d3b74e4f5797. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.566426] env[61907]: DEBUG nova.network.neutron [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "address": "fa:16:3e:4a:e6:f4", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404d6d57-b8", "ovs_interfaceid": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.655882] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 971.684360] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 971.684642] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 971.684811] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 971.685012] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 971.685802] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 971.685976] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 971.686346] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 971.686427] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 971.686602] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 971.686771] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 971.686955] env[61907]: DEBUG nova.virt.hardware [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 971.687832] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46756ca5-8951-41f6-b2da-9ec828d35a75 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.697858] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8942faa1-163f-4b6c-bca8-93379c602709 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.739582] env[61907]: DEBUG oslo_vmware.api [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244202, 'name': PowerOnVM_Task, 'duration_secs': 0.442149} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.740751] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 971.740751] env[61907]: INFO nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Took 8.80 seconds to spawn the instance on the hypervisor. [ 971.740751] env[61907]: DEBUG nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 971.741063] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdd1dc6-9bd0-4ad1-a844-e91c5ba7cb24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.753416] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.755328] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Instance network_info: |[{"id": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "address": "fa:16:3e:f2:c0:21", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped422e68-2f", "ovs_interfaceid": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 971.755328] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:c0:21', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ed422e68-2fc3-4015-8d2a-716f55d3ea41', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.762448] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Creating folder: Project (8c30836c7053443aaab6e1dbf7783259). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.765805] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c0fe3fc-d3fc-4595-a228-0e74ac321675 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.778209] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Created folder: Project (8c30836c7053443aaab6e1dbf7783259) in parent group-v268168. [ 971.778209] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Creating folder: Instances. Parent ref: group-v268301. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 971.778209] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-430e8dfc-9286-4ebd-89d3-978ca5802d01 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.791891] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244203, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.793119] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Created folder: Instances in parent group-v268301. [ 971.793368] env[61907]: DEBUG oslo.service.loopingcall [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.793592] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.794158] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-71004174-28d5-4b45-9e50-dd4372a108ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.810908] env[61907]: DEBUG nova.network.neutron [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [{"id": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "address": "fa:16:3e:e6:bb:60", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b42cc64-10", "ovs_interfaceid": "8b42cc64-101e-4eec-953d-5c2cc55dd252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.824012] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528af506-c586-f45d-d4b2-c2427d29d00e, 'name': SearchDatastore_Task, 'duration_secs': 0.010558} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.825624] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.825896] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/b43efe48-95de-46ec-8cbb-c24cf7bd68a1.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 971.826633] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.826633] env[61907]: value = "task-1244206" [ 971.826633] env[61907]: _type = "Task" [ 971.826633] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.826850] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f068f81f-af9d-4ca8-8f20-8f9b173e71b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.839702] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244206, 'name': CreateVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.841227] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 971.841227] env[61907]: value = "task-1244207" [ 971.841227] env[61907]: _type = "Task" [ 971.841227] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.849786] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244207, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.881127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0c9df068-811d-4aee-b08b-59fbcdb0af57 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.192s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.904855] env[61907]: DEBUG nova.compute.manager [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Received event network-vif-plugged-ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 971.905126] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Acquiring lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.905369] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.905549] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.905723] env[61907]: DEBUG nova.compute.manager [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] No waiting events found dispatching network-vif-plugged-ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 971.905890] env[61907]: WARNING nova.compute.manager [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Received unexpected event network-vif-plugged-ed422e68-2fc3-4015-8d2a-716f55d3ea41 for instance with vm_state building and task_state spawning. [ 971.906386] env[61907]: DEBUG nova.compute.manager [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Received event network-changed-ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 971.906612] env[61907]: DEBUG nova.compute.manager [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Refreshing instance network info cache due to event network-changed-ed422e68-2fc3-4015-8d2a-716f55d3ea41. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 971.906820] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Acquiring lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.906951] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Acquired lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.907125] env[61907]: DEBUG nova.network.neutron [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Refreshing network info cache for port ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 971.975703] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Successfully updated port: bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 972.068335] env[61907]: DEBUG oslo_concurrency.lockutils [req-320def59-b716-41ac-879c-4a7fc342156c req-7c3aeddd-628a-4d69-8323-1ef309582fb9 service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.261378] env[61907]: INFO nova.compute.manager [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Took 19.70 seconds to build instance. [ 972.288367] env[61907]: DEBUG oslo_vmware.api [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244203, 'name': PowerOnVM_Task, 'duration_secs': 0.593135} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.288657] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 972.288856] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6e31beee-dc00-4924-b0b6-6224a1576308 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance '5475c612-e718-49eb-9760-9cfedbd7931d' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 972.313652] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-f87e45da-187b-4aad-b7bf-b4228a8b2f1f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.313931] env[61907]: DEBUG nova.objects.instance [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'migration_context' on Instance uuid f87e45da-187b-4aad-b7bf-b4228a8b2f1f {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.342881] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244206, 'name': CreateVM_Task, 'duration_secs': 0.423976} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.346958] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 972.347870] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.348144] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.348587] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 972.349322] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3511760-895e-4c18-88d2-a97e44f6dff9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.356253] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244207, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.46961} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.357032] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/b43efe48-95de-46ec-8cbb-c24cf7bd68a1.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 972.357342] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 972.357666] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb290197-63f4-4de1-8aa8-f7532795dcd7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.361314] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 972.361314] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528afa68-446f-0796-46fe-3597ad8d16c6" [ 972.361314] env[61907]: _type = "Task" [ 972.361314] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.367040] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 972.367040] env[61907]: value = "task-1244208" [ 972.367040] env[61907]: _type = "Task" [ 972.367040] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.374725] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528afa68-446f-0796-46fe-3597ad8d16c6, 'name': SearchDatastore_Task, 'duration_secs': 0.00893} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.375310] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.375615] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 972.375887] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.376034] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.376152] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 972.376415] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3719d9dd-11bb-4767-ab86-af6d9982ec37 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.382421] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244208, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.388553] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 972.388765] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 972.389471] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2335225b-1f48-4f8f-86e2-fc3dd8d0df35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.394423] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 972.394423] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c2ddaa-da81-fa6e-0e9e-4f6b197466ea" [ 972.394423] env[61907]: _type = "Task" [ 972.394423] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.402900] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c2ddaa-da81-fa6e-0e9e-4f6b197466ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.477799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.477960] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 972.478124] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 972.508443] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.508773] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 972.615135] env[61907]: DEBUG nova.network.neutron [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updated VIF entry in instance network info cache for port ed422e68-2fc3-4015-8d2a-716f55d3ea41. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 972.615794] env[61907]: DEBUG nova.network.neutron [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [{"id": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "address": "fa:16:3e:f2:c0:21", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped422e68-2f", "ovs_interfaceid": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.701503] env[61907]: DEBUG nova.compute.manager [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Received event network-vif-plugged-bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 972.701819] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Acquiring lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.702057] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.702241] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.702457] env[61907]: DEBUG nova.compute.manager [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] No waiting events found dispatching network-vif-plugged-bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 972.702645] env[61907]: WARNING nova.compute.manager [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Received unexpected event network-vif-plugged-bd27cf86-5b58-48e0-b10f-ae391a48f9f2 for instance with vm_state building and task_state spawning. [ 972.702813] env[61907]: DEBUG nova.compute.manager [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Received event network-changed-bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 972.702974] env[61907]: DEBUG nova.compute.manager [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Refreshing instance network info cache due to event network-changed-bd27cf86-5b58-48e0-b10f-ae391a48f9f2. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 972.703513] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Acquiring lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 972.763913] env[61907]: DEBUG oslo_concurrency.lockutils [None req-865ed5c6-6bc6-4740-8515-119f1b5c8767 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.214s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.816765] env[61907]: DEBUG nova.objects.base [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 972.817457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b59cd5c-578c-4827-958c-2af18c40a632 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.837462] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c18124c-9592-4189-8f38-ed7c78bf1ced {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.842944] env[61907]: DEBUG oslo_vmware.api [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 972.842944] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c8aa20-4d52-bd3b-75bf-23050743337e" [ 972.842944] env[61907]: _type = "Task" [ 972.842944] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.851053] env[61907]: DEBUG oslo_vmware.api [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c8aa20-4d52-bd3b-75bf-23050743337e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.875556] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244208, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068901} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.875818] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.876607] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20ec21cb-e012-448c-8f13-7f1360624ba0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.899183] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/b43efe48-95de-46ec-8cbb-c24cf7bd68a1.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.899531] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-203b1591-c387-43fd-8caf-86aefe1fa842 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.923653] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c2ddaa-da81-fa6e-0e9e-4f6b197466ea, 'name': SearchDatastore_Task, 'duration_secs': 0.008779} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.925622] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 972.925622] env[61907]: value = "task-1244209" [ 972.925622] env[61907]: _type = "Task" [ 972.925622] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.925831] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdd0570d-a8c9-48c1-8a65-bb7adc068c5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.936584] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244209, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.937898] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 972.937898] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fab42b-816b-e313-0636-e12128012f51" [ 972.937898] env[61907]: _type = "Task" [ 972.937898] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.945793] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fab42b-816b-e313-0636-e12128012f51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.024211] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 973.119527] env[61907]: DEBUG oslo_concurrency.lockutils [req-bbb514a2-9829-489a-a202-d11a4bffadca req-dce89239-dd9a-4466-8a78-a7b21050e21d service nova] Releasing lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.354668] env[61907]: DEBUG oslo_vmware.api [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c8aa20-4d52-bd3b-75bf-23050743337e, 'name': SearchDatastore_Task, 'duration_secs': 0.006765} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.354974] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.355327] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.400514] env[61907]: DEBUG nova.network.neutron [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Updating instance_info_cache with network_info: [{"id": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "address": "fa:16:3e:c5:e6:95", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd27cf86-5b", "ovs_interfaceid": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.437255] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244209, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.447713] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fab42b-816b-e313-0636-e12128012f51, 'name': SearchDatastore_Task, 'duration_secs': 0.012359} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.447910] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.448197] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa/f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.448719] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6630c33-4028-4b05-8fdf-77723dccfa7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.455301] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 973.455301] env[61907]: value = "task-1244210" [ 973.455301] env[61907]: _type = "Task" [ 973.455301] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.462552] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244210, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.469015] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.469255] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.689729] env[61907]: DEBUG nova.compute.manager [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 973.691230] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90811598-9a7f-4dfd-aaa2-6b43f996d539 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.904571] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.905172] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance network_info: |[{"id": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "address": "fa:16:3e:c5:e6:95", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd27cf86-5b", "ovs_interfaceid": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 973.905557] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Acquired lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 973.905791] env[61907]: DEBUG nova.network.neutron [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Refreshing network info cache for port bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 973.906966] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e6:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd27cf86-5b58-48e0-b10f-ae391a48f9f2', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 973.914871] env[61907]: DEBUG oslo.service.loopingcall [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.920560] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 973.921418] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f8cb3b2-04fc-4af2-b6db-61ab68533bd9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.718144] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 974.718605] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 974.720027] env[61907]: INFO nova.compute.manager [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] instance snapshotting [ 974.728495] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42aadbb3-c82a-4856-b5e2-677f93288675 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.734020] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a6e58b-b95f-4294-b91d-78ce9661c795 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.734838] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 974.734838] env[61907]: value = "task-1244211" [ 974.734838] env[61907]: _type = "Task" [ 974.734838] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.741417] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244209, 'name': ReconfigVM_Task, 'duration_secs': 0.696525} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.746432] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/b43efe48-95de-46ec-8cbb-c24cf7bd68a1.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.773941] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-025903a0-ff0c-424f-b5b2-711b2c950fab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.775940] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244210, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484819} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.779911] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e51e903-850f-4f73-a38c-b1f5f0d5d96b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.782532] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa/f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 974.782775] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 974.783588] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e85f3b2-8ec2-4806-ad30-150b67093940 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.794053] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0031910b-110e-4691-ad37-5b39ca79b87c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.796197] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244211, 'name': CreateVM_Task} progress is 15%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.799421] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 974.799421] env[61907]: value = "task-1244212" [ 974.799421] env[61907]: _type = "Task" [ 974.799421] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.824010] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfiguring VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 974.834915] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d94b83c-300d-45f0-a72a-f35cda83805b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.849798] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 974.849798] env[61907]: value = "task-1244213" [ 974.849798] env[61907]: _type = "Task" [ 974.849798] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.860243] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244212, 'name': Rename_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.862913] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 974.862913] env[61907]: value = "task-1244214" [ 974.862913] env[61907]: _type = "Task" [ 974.862913] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.867566] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244213, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.874679] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.958280] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7a9b98-a51d-4126-8064-16b7a8d4fa20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.965846] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4512312-2702-43cc-8649-a309edf1e5cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.996043] env[61907]: DEBUG nova.network.neutron [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Updated VIF entry in instance network info cache for port bd27cf86-5b58-48e0-b10f-ae391a48f9f2. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 974.996455] env[61907]: DEBUG nova.network.neutron [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Updating instance_info_cache with network_info: [{"id": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "address": "fa:16:3e:c5:e6:95", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbd27cf86-5b", "ovs_interfaceid": "bd27cf86-5b58-48e0-b10f-ae391a48f9f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.999055] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3eafcf-8123-49e8-a4c5-ba56a9c91b95 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.005697] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd7aa4d-601d-4535-bdc6-7854fe04a63d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.020562] env[61907]: DEBUG nova.compute.provider_tree [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.236482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.236482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.236639] env[61907]: DEBUG nova.compute.manager [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Going to confirm migration 3 {{(pid=61907) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 975.251353] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244211, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.328814] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244212, 'name': Rename_Task, 'duration_secs': 0.265956} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.328814] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.329065] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8f789dbe-61d4-48cb-b400-df0e559f9e25 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.335118] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 975.335118] env[61907]: value = "task-1244215" [ 975.335118] env[61907]: _type = "Task" [ 975.335118] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.342118] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.354254] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 975.357341] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-61bdc744-e630-4911-88ad-7c08300627ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.363575] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244213, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.186165} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.367127] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 975.367463] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 975.367463] env[61907]: value = "task-1244216" [ 975.367463] env[61907]: _type = "Task" [ 975.367463] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.368120] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01263ea8-d85a-4a89-a6e3-393442dced18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.376281] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.394714] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa/f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 975.398099] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15a36f6b-cf08-4676-8f55-cb61d1f0e6f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.411989] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244216, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.417055] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 975.417055] env[61907]: value = "task-1244217" [ 975.417055] env[61907]: _type = "Task" [ 975.417055] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.425911] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244217, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.501965] env[61907]: DEBUG oslo_concurrency.lockutils [req-e5c8eca6-d1b5-4884-aa5b-24b5372ec70d req-e7acfb50-4382-4115-beb7-673fca8f46ec service nova] Releasing lock "refresh_cache-4430db3b-0276-42b6-a67e-386a164ddc0e" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 975.513736] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 975.513930] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 975.543593] env[61907]: ERROR nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [req-1d695121-f1c6-44d9-a914-9449bb3ca0ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-1d695121-f1c6-44d9-a914-9449bb3ca0ef"}]} [ 975.561483] env[61907]: DEBUG nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 975.576033] env[61907]: DEBUG nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 975.576321] env[61907]: DEBUG nova.compute.provider_tree [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.587617] env[61907]: DEBUG nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 975.606010] env[61907]: DEBUG nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 975.755682] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244211, 'name': CreateVM_Task, 'duration_secs': 0.733192} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.756058] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 975.756814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.756942] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.758043] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 975.758043] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00efdd71-c213-4ac2-9a3e-c2b6f6834f59 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.764250] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 975.764250] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9a4f-e7e5-d066-e3dd-57ade32018bc" [ 975.764250] env[61907]: _type = "Task" [ 975.764250] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.773415] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9a4f-e7e5-d066-e3dd-57ade32018bc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.806560] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.806796] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquired lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.806986] env[61907]: DEBUG nova.network.neutron [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.807205] env[61907]: DEBUG nova.objects.instance [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'info_cache' on Instance uuid 5475c612-e718-49eb-9760-9cfedbd7931d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.812974] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebc20c8-ee4f-4e7a-b39d-fda0ea240741 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.819416] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-577f85e7-7dc6-4edb-b929-cd1c88d3d2e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.855332] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7fc69b-1d5c-444f-80c7-f4c8da33a361 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.862695] env[61907]: DEBUG oslo_vmware.api [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244215, 'name': PowerOnVM_Task, 'duration_secs': 0.525799} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.864850] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.865138] env[61907]: INFO nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Took 8.88 seconds to spawn the instance on the hypervisor. [ 975.865374] env[61907]: DEBUG nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 975.866215] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f17d31d4-4fda-4c64-b068-5c7ab96e3de4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.871879] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7c63f3-174e-4b29-82fd-e06d0b46d992 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.894865] env[61907]: DEBUG nova.compute.provider_tree [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 975.896290] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.900653] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244216, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.926879] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244217, 'name': ReconfigVM_Task, 'duration_secs': 0.382386} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.927175] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Reconfigured VM instance instance-0000005b to attach disk [datastore2] f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa/f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.927786] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0d6152e-5d7c-4b74-89ae-92efd91354dd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.933457] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 975.933457] env[61907]: value = "task-1244218" [ 975.933457] env[61907]: _type = "Task" [ 975.933457] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.940956] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244218, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.046680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.046919] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.047140] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 976.274975] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9a4f-e7e5-d066-e3dd-57ade32018bc, 'name': SearchDatastore_Task, 'duration_secs': 0.01215} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.275333] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 976.275618] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 976.275887] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.276081] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.276335] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 976.276598] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2ada620-a486-4b55-ae2c-d0818f14646e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.287137] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 976.287324] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 976.288027] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29e5f6aa-3ea4-47f8-95cd-55814f20b0ca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.292907] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 976.292907] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b74763-30a1-a322-0132-57cc92345b93" [ 976.292907] env[61907]: _type = "Task" [ 976.292907] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.303923] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b74763-30a1-a322-0132-57cc92345b93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.377592] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.385376] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244216, 'name': CreateSnapshot_Task, 'duration_secs': 0.795244} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.385652] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 976.386383] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87022f13-2f9c-4d8d-99dd-87fe8c64dca5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.405738] env[61907]: INFO nova.compute.manager [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Took 21.29 seconds to build instance. [ 976.428102] env[61907]: DEBUG nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 115 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 976.428369] env[61907]: DEBUG nova.compute.provider_tree [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 115 to 116 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 976.428554] env[61907]: DEBUG nova.compute.provider_tree [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 976.442290] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244218, 'name': Rename_Task, 'duration_secs': 0.181957} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.443024] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.443294] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-81e58524-94a3-4ed9-90e8-05212391243d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.448907] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 976.448907] env[61907]: value = "task-1244219" [ 976.448907] env[61907]: _type = "Task" [ 976.448907] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.456467] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244219, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.800070] env[61907]: INFO nova.compute.manager [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Rescuing [ 976.800380] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.800587] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.800811] env[61907]: DEBUG nova.network.neutron [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 976.807648] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b74763-30a1-a322-0132-57cc92345b93, 'name': SearchDatastore_Task, 'duration_secs': 0.02908} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.808433] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e8a2bb9-daea-4cfb-a9c6-97c07f529e03 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.816286] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 976.816286] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5281406e-907b-d691-d7fc-c70402e8a882" [ 976.816286] env[61907]: _type = "Task" [ 976.816286] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.826746] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5281406e-907b-d691-d7fc-c70402e8a882, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.877697] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.902910] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 976.903272] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-425829cb-99ec-456f-ac9a-8fcf10a5dd06 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.907842] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7476df75-3ad7-468a-aed8-6608630a0529 tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.804s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.910965] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 976.910965] env[61907]: value = "task-1244220" [ 976.910965] env[61907]: _type = "Task" [ 976.910965] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.920228] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244220, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.959704] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244219, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.009019] env[61907]: DEBUG nova.network.neutron [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [{"id": "e44f586e-da28-4e87-95ea-518bbf489218", "address": "fa:16:3e:33:d4:e3", "network": {"id": "3e6409b6-41a1-4465-b8cb-3ba3c52052f2", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-63281994-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d8b341077bf045e0b079bbe8aa08cae0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1fb81f98-6f5a-47ab-a512-27277591d064", "external-id": "nsx-vlan-transportzone-624", "segmentation_id": 624, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44f586e-da", "ovs_interfaceid": "e44f586e-da28-4e87-95ea-518bbf489218", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.257992] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [{"id": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "address": "fa:16:3e:80:83:7b", "network": {"id": "c5da93e7-e079-40d9-a1d7-3496f5d01771", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1601373482-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "89e4f90a5fe44853a926ceba2f5150dd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d19577c9-1b2e-490b-8031-2f278dd3f570", "external-id": "nsx-vlan-transportzone-611", "segmentation_id": 611, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9a182520-96", "ovs_interfaceid": "9a182520-96f9-434d-844a-fe3e6cc88fbc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.328681] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5281406e-907b-d691-d7fc-c70402e8a882, 'name': SearchDatastore_Task, 'duration_secs': 0.01425} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.328954] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.329234] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 977.329499] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8f6d12c-7121-46b2-80c6-537fe45ca82c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.336329] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 977.336329] env[61907]: value = "task-1244221" [ 977.336329] env[61907]: _type = "Task" [ 977.336329] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.345909] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244221, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.378630] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.424929] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244220, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.441070] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 4.085s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.460567] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244219, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.512840] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Releasing lock "refresh_cache-5475c612-e718-49eb-9760-9cfedbd7931d" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.513038] env[61907]: DEBUG nova.objects.instance [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lazy-loading 'migration_context' on Instance uuid 5475c612-e718-49eb-9760-9cfedbd7931d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.537055] env[61907]: DEBUG nova.network.neutron [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.760641] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-2ab93ab9-bf69-4525-8df6-eef83dd24bc1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.761051] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 977.761260] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.761615] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.761882] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.762086] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.762280] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 977.762487] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 977.848072] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244221, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487705} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.848396] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 977.848562] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 977.848815] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d74ed394-db9b-4b18-a74e-f7f3189b4c8c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.854572] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 977.854572] env[61907]: value = "task-1244222" [ 977.854572] env[61907]: _type = "Task" [ 977.854572] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.861874] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244222, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.875805] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.920159] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244220, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.960689] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244219, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.013547] env[61907]: INFO nova.scheduler.client.report [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocation for migration 24370906-b77f-4fe5-9015-61f8ba73b556 [ 978.018245] env[61907]: DEBUG nova.objects.base [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Object Instance<5475c612-e718-49eb-9760-9cfedbd7931d> lazy-loaded attributes: info_cache,migration_context {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 978.019253] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe1cdd78-930a-4707-865f-8887b36a8e80 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.040380] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.042308] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e01812b4-2298-4d04-9d35-9ee21d197472 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.048047] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 978.048047] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528a95bc-949d-34d0-6605-d9d3b6be0d2a" [ 978.048047] env[61907]: _type = "Task" [ 978.048047] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.057235] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528a95bc-949d-34d0-6605-d9d3b6be0d2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.265479] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.265719] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.265888] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.266126] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 978.266945] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4035abb7-48ec-46dd-9f31-15eb372c456b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.275210] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0323695-409d-4e70-8026-821fb82d3fe6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.288771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a01d000-f8dc-4206-a514-26fb2a8cbfc0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.294986] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28742525-6875-40f0-9d68-d4b5123bfba2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.340993] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179608MB free_disk=147GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 978.341318] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.341636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.367537] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244222, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0725} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.367936] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 978.369217] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e83a8a-f221-4a3d-b710-ae5a496e5b8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.410069] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 978.417539] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45501c95-d41e-4f22-a34f-c5a94cd8d7e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.441587] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.456128] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244220, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.461816] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 978.461816] env[61907]: value = "task-1244223" [ 978.461816] env[61907]: _type = "Task" [ 978.461816] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.471599] env[61907]: DEBUG oslo_vmware.api [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244219, 'name': PowerOnVM_Task, 'duration_secs': 1.980602} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.472457] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.472772] env[61907]: INFO nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Took 9.18 seconds to spawn the instance on the hypervisor. [ 978.473144] env[61907]: DEBUG nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 978.474259] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f8a3b56-ce8a-471a-a051-6ad678fd7b8c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.482659] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244223, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.520531] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e6315ef-1d08-4c13-b061-7bd39bf04f6b tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.720s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.560272] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528a95bc-949d-34d0-6605-d9d3b6be0d2a, 'name': SearchDatastore_Task, 'duration_secs': 0.008144} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.560628] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.732446] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.733014] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.733147] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.733387] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.733609] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.736334] env[61907]: INFO nova.compute.manager [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Terminating instance [ 978.878817] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.950855] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244220, 'name': CloneVM_Task, 'duration_secs': 1.974653} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.951150] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Created linked-clone VM from snapshot [ 978.951905] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9d432c-e2af-442f-b7fb-08ce8f360cb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.958947] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Uploading image e4d47f58-1099-4861-9fe7-673ca090fbec {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 978.970225] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244223, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.971632] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 978.971890] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-907d5928-163e-48d4-b0d8-b8e5f77ea458 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.977871] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 978.977871] env[61907]: value = "task-1244224" [ 978.977871] env[61907]: _type = "Task" [ 978.977871] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.988090] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244224, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.999645] env[61907]: INFO nova.compute.manager [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Took 21.30 seconds to build instance. [ 979.075198] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.075610] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bcef713-0f96-4830-bf75-db5a9dfd07a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.083297] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 979.083297] env[61907]: value = "task-1244225" [ 979.083297] env[61907]: _type = "Task" [ 979.083297] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.094489] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244225, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.214035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "696c8d63-5268-4239-8d41-079e9f3f7c66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.214035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.240160] env[61907]: DEBUG nova.compute.manager [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 979.240415] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 979.241567] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0032212c-479f-41e8-88dc-a1364750c51d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.250847] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.251156] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d92e75e-e787-45d2-819a-ab9a841f04cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.257730] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 979.257730] env[61907]: value = "task-1244226" [ 979.257730] env[61907]: _type = "Task" [ 979.257730] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.267983] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244226, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.354070] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Applying migration context for instance 5475c612-e718-49eb-9760-9cfedbd7931d as it has an incoming, in-progress migration 10a7713a-f8c3-42d0-b6ef-9f707de1f035. Migration status is finished {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 979.355535] env[61907]: INFO nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating resource usage from migration 10a7713a-f8c3-42d0-b6ef-9f707de1f035 [ 979.379633] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.382463] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.382610] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.382734] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 23526967-cd8f-4581-b9c5-1c270d385163 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.382853] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.382967] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 53176c9e-d15c-49d5-b4a9-22b780279ecb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383091] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance fcde07bf-858a-4377-b27c-4f17356306af actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383223] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f50bf687-609c-44e1-bd75-cf4efaeadb71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383345] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f87e45da-187b-4aad-b7bf-b4228a8b2f1f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383458] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Migration 10a7713a-f8c3-42d0-b6ef-9f707de1f035 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 979.383568] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5475c612-e718-49eb-9760-9cfedbd7931d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383677] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 6a092752-058f-4196-848e-58a94d084cc7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383786] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance b43efe48-95de-46ec-8cbb-c24cf7bd68a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.383893] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.384122] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 4430db3b-0276-42b6-a67e-386a164ddc0e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 979.473717] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244223, 'name': ReconfigVM_Task, 'duration_secs': 0.678603} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.474084] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Reconfigured VM instance instance-0000005c to attach disk [datastore2] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 979.474716] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f256fab-aa7e-4495-af29-10ddfa2e869b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.483392] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 979.483392] env[61907]: value = "task-1244227" [ 979.483392] env[61907]: _type = "Task" [ 979.483392] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.489904] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244224, 'name': Destroy_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.494479] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244227, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.502089] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d07abd24-de38-41bd-94ff-ef2aba80e4b0 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.814s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.593610] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244225, 'name': PowerOffVM_Task, 'duration_secs': 0.258597} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.593879] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.594943] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55bc9cc7-5b59-41e6-9d07-9ce68654c28b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.612127] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a50b8bcf-75bb-48f2-b534-0ee894036bb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.642032] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.642360] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc30b91a-7914-4402-9ba5-01e5b479cc53 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.648709] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 979.648709] env[61907]: value = "task-1244228" [ 979.648709] env[61907]: _type = "Task" [ 979.648709] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.658599] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 979.658831] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.659097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.659257] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.659445] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.659690] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bf463f00-30db-4595-802e-537fad455ec5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.667913] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.668051] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.668815] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c02b7a4f-0f84-477a-bb1c-8b37f3b35152 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.675586] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 979.675586] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c0b-d429-b38f-0a02-8e29c264c5db" [ 979.675586] env[61907]: _type = "Task" [ 979.675586] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.683528] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c0b-d429-b38f-0a02-8e29c264c5db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.688831] env[61907]: DEBUG nova.compute.manager [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Received event network-changed-ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 979.689077] env[61907]: DEBUG nova.compute.manager [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Refreshing instance network info cache due to event network-changed-ed422e68-2fc3-4015-8d2a-716f55d3ea41. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 979.689314] env[61907]: DEBUG oslo_concurrency.lockutils [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] Acquiring lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.689568] env[61907]: DEBUG oslo_concurrency.lockutils [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] Acquired lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.689804] env[61907]: DEBUG nova.network.neutron [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Refreshing network info cache for port ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.716026] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 979.768472] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244226, 'name': PowerOffVM_Task, 'duration_secs': 0.306517} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.768788] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 979.768998] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 979.769351] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e79b6285-a65e-4911-8c24-8efba4d98c51 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.836648] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.836887] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.837088] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleting the datastore file [datastore2] f87e45da-187b-4aad-b7bf-b4228a8b2f1f {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.837365] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9d78b52-63ff-4e5b-b9a1-ef051611fdb9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.843940] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 979.843940] env[61907]: value = "task-1244230" [ 979.843940] env[61907]: _type = "Task" [ 979.843940] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.852395] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244230, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.880201] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.890508] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 696c8d63-5268-4239-8d41-079e9f3f7c66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 979.890785] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 979.890945] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3328MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 979.992817] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244224, 'name': Destroy_Task, 'duration_secs': 0.731894} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.993575] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Destroyed the VM [ 979.993911] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 979.994247] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ba81d044-ed4c-4325-a083-ae71761f3eca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.999627] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244227, 'name': Rename_Task, 'duration_secs': 0.327523} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.000563] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 980.000889] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ab68b735-4065-4848-aacd-37f618bf2817 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.006341] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 980.006341] env[61907]: value = "task-1244231" [ 980.006341] env[61907]: _type = "Task" [ 980.006341] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.010779] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 980.010779] env[61907]: value = "task-1244232" [ 980.010779] env[61907]: _type = "Task" [ 980.010779] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.017719] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244231, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.029187] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244232, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.121560] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54d01296-f4ff-4709-b770-45f95a232487 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.129344] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a7cb5f5-e07b-492f-8f9e-8205d21e7923 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.159580] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6158ae0c-97f9-472a-b4b3-40691ac72f56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.167293] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55bcf2b-f371-47f1-ba85-8535d943a480 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.180554] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.190423] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52975c0b-d429-b38f-0a02-8e29c264c5db, 'name': SearchDatastore_Task, 'duration_secs': 0.010641} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.193572] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72d9bd6b-55de-4239-8c6d-2be024cda74d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.200158] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 980.200158] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d95287-a394-06c9-d473-414fa48529ce" [ 980.200158] env[61907]: _type = "Task" [ 980.200158] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.207982] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d95287-a394-06c9-d473-414fa48529ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.235554] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.353549] env[61907]: DEBUG oslo_vmware.api [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244230, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.166667} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.353876] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 980.354095] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 980.354281] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 980.354495] env[61907]: INFO nova.compute.manager [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Took 1.11 seconds to destroy the instance on the hypervisor. [ 980.354743] env[61907]: DEBUG oslo.service.loopingcall [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 980.354967] env[61907]: DEBUG nova.compute.manager [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 980.355080] env[61907]: DEBUG nova.network.neutron [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 980.381689] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.413030] env[61907]: DEBUG nova.network.neutron [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updated VIF entry in instance network info cache for port ed422e68-2fc3-4015-8d2a-716f55d3ea41. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.413373] env[61907]: DEBUG nova.network.neutron [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [{"id": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "address": "fa:16:3e:f2:c0:21", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped422e68-2f", "ovs_interfaceid": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.518776] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244231, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.524644] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244232, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.686672] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 980.711113] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d95287-a394-06c9-d473-414fa48529ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009577} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.711396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.711657] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. {{(pid=61907) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 980.711910] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffaf9475-671e-41cb-9549-4aba28b36938 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.718243] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 980.718243] env[61907]: value = "task-1244233" [ 980.718243] env[61907]: _type = "Task" [ 980.718243] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.727127] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.883908] env[61907]: DEBUG oslo_vmware.api [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244214, 'name': ReconfigVM_Task, 'duration_secs': 5.819247} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.884664] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.885098] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Reconfigured VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 980.918563] env[61907]: DEBUG oslo_concurrency.lockutils [req-246edc2a-9eeb-4312-bc48-2080bb627b28 req-73e7d373-97bf-4425-8c9d-7fc82888c446 service nova] Releasing lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.020482] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244231, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.029846] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244232, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.167487] env[61907]: DEBUG nova.network.neutron [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.191038] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 981.191630] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.850s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.191793] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.631s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.228503] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244233, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473226} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.228778] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk. [ 981.229588] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b919d1a-574a-4885-a9b8-be8e69ae586d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.255229] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 981.255526] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97fd9a1a-0fdf-4894-8eb7-e941111d87be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.277682] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 981.277682] env[61907]: value = "task-1244234" [ 981.277682] env[61907]: _type = "Task" [ 981.277682] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.288147] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.520284] env[61907]: DEBUG oslo_vmware.api [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244231, 'name': RemoveSnapshot_Task, 'duration_secs': 1.04881} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.521073] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 981.526715] env[61907]: DEBUG oslo_vmware.api [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244232, 'name': PowerOnVM_Task, 'duration_secs': 1.292346} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.527226] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 981.527535] env[61907]: INFO nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Took 9.87 seconds to spawn the instance on the hypervisor. [ 981.527873] env[61907]: DEBUG nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 981.528631] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9208312-40ed-456d-9c62-ddf79ce57df5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.670548] env[61907]: INFO nova.compute.manager [-] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Took 1.32 seconds to deallocate network for instance. [ 981.785817] env[61907]: DEBUG nova.compute.manager [req-18b34008-6843-4039-9b39-46cde43e1e6c req-f6982161-b123-436d-b7d1-713685d23df8 service nova] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Received event network-vif-deleted-8b42cc64-101e-4eec-953d-5c2cc55dd252 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 981.792018] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244234, 'name': ReconfigVM_Task, 'duration_secs': 0.315866} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.792562] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57-rescue.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.793552] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44caeded-f232-44e3-bb3f-9c4195f07aef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.823462] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-905c5838-3c9d-499c-8212-c43e108eb253 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.838708] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 981.838708] env[61907]: value = "task-1244235" [ 981.838708] env[61907]: _type = "Task" [ 981.838708] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.846774] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244235, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.921928] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9114988e-703e-4766-bc02-5c419b9b03ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.929674] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd7fafb5-3a15-4012-8e1a-5eabb50757ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.960840] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 981.962500] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10d5b0a0-366e-4a79-8d15-c532181143d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.969705] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4eb3643-437d-4e6e-9e87-85d544efc6cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.982239] env[61907]: DEBUG nova.compute.provider_tree [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.029045] env[61907]: WARNING nova.compute.manager [None req-6512290a-2102-4140-9328-20b85a866803 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Image not found during snapshot: nova.exception.ImageNotFound: Image e4d47f58-1099-4861-9fe7-673ca090fbec could not be found. [ 982.043621] env[61907]: INFO nova.compute.manager [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Took 21.34 seconds to build instance. [ 982.180859] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.266054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "6a092752-058f-4196-848e-58a94d084cc7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.266054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.266054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "6a092752-058f-4196-848e-58a94d084cc7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.266054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.266054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.268378] env[61907]: INFO nova.compute.manager [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Terminating instance [ 982.349082] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244235, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.383037] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.383343] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.383529] env[61907]: DEBUG nova.network.neutron [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.484568] env[61907]: DEBUG nova.scheduler.client.report [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 982.545809] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fc466a99-9cb5-4720-9504-7c603b809fa9 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.850s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.772418] env[61907]: DEBUG nova.compute.manager [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 982.772673] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.774130] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cbf5859-b3f3-435c-951f-1a92769c8d7a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.781403] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.781632] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7c342ac-1362-4ebe-a504-10afe0e1d146 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.786965] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 982.786965] env[61907]: value = "task-1244236" [ 982.786965] env[61907]: _type = "Task" [ 982.786965] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.794283] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244236, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.848466] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244235, 'name': ReconfigVM_Task, 'duration_secs': 0.870745} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.848775] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 982.849089] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-309b81d8-b39a-477d-b94b-cc7a8cf6a627 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.855517] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 982.855517] env[61907]: value = "task-1244237" [ 982.855517] env[61907]: _type = "Task" [ 982.855517] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.863211] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.974327] env[61907]: INFO nova.compute.manager [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Rebuilding instance [ 983.021023] env[61907]: DEBUG nova.compute.manager [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.021023] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13b4194e-feb3-4579-b84e-6d7a93a63aac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.299729] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244236, 'name': PowerOffVM_Task, 'duration_secs': 0.230803} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.300212] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.300564] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.301209] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ca21413-ea1f-4b87-8e6c-b48e809feb41 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.362678] env[61907]: INFO nova.network.neutron [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Port 404d6d57-b8d5-42c7-b255-d3b74e4f5797 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 983.363622] env[61907]: DEBUG nova.network.neutron [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.376174] env[61907]: DEBUG oslo_vmware.api [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244237, 'name': PowerOnVM_Task, 'duration_secs': 0.442151} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.376174] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 983.378742] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.379274] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.379620] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleting the datastore file [datastore2] 6a092752-058f-4196-848e-58a94d084cc7 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.380962] env[61907]: DEBUG nova.compute.manager [None req-d9033216-559b-45c3-ade1-90b59ae1420e tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 983.381819] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-253487ac-dab3-48fc-a380-7406fb3e5ca6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.386829] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24cfb16b-a91c-4531-a0fe-d46b0f968bdf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.399076] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for the task: (returnval){ [ 983.399076] env[61907]: value = "task-1244239" [ 983.399076] env[61907]: _type = "Task" [ 983.399076] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.413702] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244239, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.499022] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.305s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.500140] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.265s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.502565] env[61907]: INFO nova.compute.claims [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.873819] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.914596] env[61907]: DEBUG oslo_vmware.api [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Task: {'id': task-1244239, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.35468} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.916216] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 983.916216] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 983.916216] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 983.916216] env[61907]: INFO nova.compute.manager [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Took 1.14 seconds to destroy the instance on the hypervisor. [ 983.916216] env[61907]: DEBUG oslo.service.loopingcall [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 983.916607] env[61907]: DEBUG nova.compute.manager [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 983.916607] env[61907]: DEBUG nova.network.neutron [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.038762] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 984.039106] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c41e40f6-6a68-46e9-a6d8-974eb0cd7a8f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.046110] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 984.046110] env[61907]: value = "task-1244240" [ 984.046110] env[61907]: _type = "Task" [ 984.046110] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.054346] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.064862] env[61907]: INFO nova.scheduler.client.report [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocation for migration 10a7713a-f8c3-42d0-b6ef-9f707de1f035 [ 984.361613] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.361930] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.362354] env[61907]: DEBUG nova.objects.instance [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'flavor' on Instance uuid f50bf687-609c-44e1-bd75-cf4efaeadb71 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 984.378646] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e6879181-ddc6-414d-9b16-c80bdc075c21 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-fcde07bf-858a-4377-b27c-4f17356306af-404d6d57-b8d5-42c7-b255-d3b74e4f5797" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.909s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.555514] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244240, 'name': PowerOffVM_Task, 'duration_secs': 0.396804} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.557741] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 984.557979] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 984.558884] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a017b4-096a-4c1e-a824-657177307919 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.566406] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 984.568504] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d93699c7-5ab6-4f01-beec-71bd248fb225 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.570631] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 9.334s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.639020] env[61907]: DEBUG nova.network.neutron [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.641040] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 984.641156] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 984.641336] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleting the datastore file [datastore2] 4430db3b-0276-42b6-a67e-386a164ddc0e {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 984.641840] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b4a2fd89-8337-476a-bef9-0e7800b02561 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.650842] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 984.650842] env[61907]: value = "task-1244242" [ 984.650842] env[61907]: _type = "Task" [ 984.650842] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.659374] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.691286] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3534456-af0a-4055-95d7-300991b94715 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.698230] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f818f41d-0d1e-4dfb-a4dd-897261441d3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.727595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81bb5df1-1e0a-4773-b22a-11eaa72e50b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.734872] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56757ef6-e209-4a07-b29f-85a6e6d82972 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.748069] env[61907]: DEBUG nova.compute.provider_tree [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.961632] env[61907]: DEBUG nova.objects.instance [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'pci_requests' on Instance uuid f50bf687-609c-44e1-bd75-cf4efaeadb71 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.143776] env[61907]: INFO nova.compute.manager [-] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Took 1.23 seconds to deallocate network for instance. [ 985.160453] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.251664] env[61907]: DEBUG nova.scheduler.client.report [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 985.458774] env[61907]: DEBUG nova.compute.manager [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 985.459232] env[61907]: DEBUG nova.compute.manager [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing instance network info cache due to event network-changed-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 985.459338] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Acquiring lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.459440] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Acquired lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.459626] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Refreshing network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 985.464182] env[61907]: DEBUG nova.objects.base [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 985.464406] env[61907]: DEBUG nova.network.neutron [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.532304] env[61907]: DEBUG nova.policy [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3163bc316cf342a487c5db97af65db60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6df6b1cd82e24d2f8aa1812575f03f8a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 985.650115] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.661713] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.58995} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.661942] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 985.662295] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 985.662504] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 985.757264] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.757927] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 985.760481] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.580s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.760709] env[61907]: DEBUG nova.objects.instance [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'resources' on Instance uuid f87e45da-187b-4aad-b7bf-b4228a8b2f1f {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.926784] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.927129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.927367] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.927564] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.927851] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.930264] env[61907]: INFO nova.compute.manager [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Terminating instance [ 986.270457] env[61907]: DEBUG nova.compute.utils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 986.270457] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 986.270457] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 986.296881] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updated VIF entry in instance network info cache for port bf950e7a-fdf3-453e-8bda-0ccca9cc8c98. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 986.297147] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [{"id": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "address": "fa:16:3e:23:ef:58", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbf950e7a-fd", "ovs_interfaceid": "bf950e7a-fdf3-453e-8bda-0ccca9cc8c98", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.312852] env[61907]: DEBUG nova.policy [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '41fc76e7ed9a46a09f96f11b28556ab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b57c1194e0f14ae498d99d302da346c5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.435105] env[61907]: DEBUG nova.compute.manager [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 986.435187] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 986.436484] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea2eef3-122e-4614-9fd4-3a7d34d2f9ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.452834] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 986.456518] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00772af6-9e02-4ddc-8fdd-a195bba7aefc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.471329] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 986.471329] env[61907]: value = "task-1244243" [ 986.471329] env[61907]: _type = "Task" [ 986.471329] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.482653] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244243, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.517062] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a016b8-f114-47ac-b21f-fc28c95ec0b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.525860] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3903b533-acb3-4e21-87f8-cd171c704ace {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.530685] env[61907]: DEBUG nova.compute.manager [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 986.530884] env[61907]: DEBUG nova.compute.manager [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing instance network info cache due to event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 986.531146] env[61907]: DEBUG oslo_concurrency.lockutils [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.531254] env[61907]: DEBUG oslo_concurrency.lockutils [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.531408] env[61907]: DEBUG nova.network.neutron [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.569557] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd09624-09ce-4eb4-8526-0680a722a03a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.578629] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-651e74d6-e65d-42a4-9188-f363f050a767 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.594444] env[61907]: DEBUG nova.compute.provider_tree [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.713680] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 986.713919] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 986.714089] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 986.714357] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 986.714438] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 986.714633] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 986.714872] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 986.715090] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 986.715305] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 986.715493] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 986.715678] env[61907]: DEBUG nova.virt.hardware [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 986.716631] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd22a19-e401-4720-86a8-f91729b47c21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.724991] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247cb864-9c6f-41d0-a0ad-a71c0fe54a3f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.741046] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:e6:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bd27cf86-5b58-48e0-b10f-ae391a48f9f2', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.748744] env[61907]: DEBUG oslo.service.loopingcall [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.749133] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 986.749505] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-394b5c7c-8913-4e35-956e-4ce1d82a05e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.768215] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.768215] env[61907]: value = "task-1244244" [ 986.768215] env[61907]: _type = "Task" [ 986.768215] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.772618] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 986.779185] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.799360] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Releasing lock "refresh_cache-fcde07bf-858a-4377-b27c-4f17356306af" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.801793] env[61907]: DEBUG nova.compute.manager [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 986.801793] env[61907]: DEBUG nova.compute.manager [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing instance network info cache due to event network-changed-0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 986.801793] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.801793] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.801793] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 986.907803] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Successfully created port: 1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.982266] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244243, 'name': PowerOffVM_Task, 'duration_secs': 0.329464} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.982569] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 986.982767] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 986.983013] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2c06dc98-9df8-40a8-8948-e719947a378f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.050057] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.050311] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.050500] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleting the datastore file [datastore2] 5475c612-e718-49eb-9760-9cfedbd7931d {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.050792] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-585c46ea-89dd-4f61-973c-31a21e2fb142 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.057747] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for the task: (returnval){ [ 987.057747] env[61907]: value = "task-1244246" [ 987.057747] env[61907]: _type = "Task" [ 987.057747] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.066175] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244246, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.097690] env[61907]: DEBUG nova.scheduler.client.report [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 987.237154] env[61907]: DEBUG nova.network.neutron [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updated VIF entry in instance network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.237531] env[61907]: DEBUG nova.network.neutron [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.282011] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.303443] env[61907]: DEBUG nova.network.neutron [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Successfully updated port: 404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.551110] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updated VIF entry in instance network info cache for port 0662d339-0c3c-49e2-9c37-6dd66f7e704e. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 987.551502] env[61907]: DEBUG nova.network.neutron [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.567341] env[61907]: DEBUG oslo_vmware.api [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Task: {'id': task-1244246, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.326976} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.568277] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.568551] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 987.568817] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 987.569095] env[61907]: INFO nova.compute.manager [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 987.569426] env[61907]: DEBUG oslo.service.loopingcall [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.569916] env[61907]: DEBUG nova.compute.manager [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 987.570079] env[61907]: DEBUG nova.network.neutron [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 987.602271] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.842s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.604593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.955s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.604865] env[61907]: DEBUG nova.objects.instance [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lazy-loading 'resources' on Instance uuid 6a092752-058f-4196-848e-58a94d084cc7 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.627883] env[61907]: INFO nova.scheduler.client.report [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocations for instance f87e45da-187b-4aad-b7bf-b4228a8b2f1f [ 987.740246] env[61907]: DEBUG oslo_concurrency.lockutils [req-5c01f61b-60f0-45ed-96e2-2ec08652df3f req-645c82fd-46ed-469e-9134-320c0bf7f731 service nova] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.779681] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.783818] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 987.807200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.822360] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.822646] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.822806] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.822997] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.823170] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.823329] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.823758] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.823953] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.824160] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.824341] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.824524] env[61907]: DEBUG nova.virt.hardware [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.825668] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cbf72a-5f5e-4552-901c-1ae96002a003 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.835681] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa1b0ed-4f7d-426d-97c0-8c6d875312f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.987272] env[61907]: DEBUG nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Received event network-vif-deleted-88b21882-cb3c-444d-8966-d16967161a8a {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 987.987430] env[61907]: DEBUG nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 987.987633] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.987905] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.988043] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.988224] env[61907]: DEBUG nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] No waiting events found dispatching network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 987.988394] env[61907]: WARNING nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received unexpected event network-vif-plugged-404d6d57-b8d5-42c7-b255-d3b74e4f5797 for instance with vm_state active and task_state None. [ 987.988560] env[61907]: DEBUG nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-changed-404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 987.988719] env[61907]: DEBUG nova.compute.manager [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing instance network info cache due to event network-changed-404d6d57-b8d5-42c7-b255-d3b74e4f5797. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 987.988884] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.053948] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f215f27-df06-45eb-9a7b-751b8edda01e req-eef0f2f6-cdfe-4dfb-9d8c-9a9e8b47c34a service nova] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.054834] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.055081] env[61907]: DEBUG nova.network.neutron [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.137777] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d84b9976-5f89-434e-963f-7a6e4d0d6984 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "f87e45da-187b-4aad-b7bf-b4228a8b2f1f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.405s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.282053] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.305733] env[61907]: DEBUG nova.network.neutron [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.346508] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aca5c761-7029-4fc8-93db-a72f0e1abc09 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.353743] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff2cf3e5-f200-4598-82df-8f698cc15b1b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.385075] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36372679-654d-4685-b4d4-c851935d80ac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.392536] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd556323-fe43-42d1-a02f-60cf016ff8e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.407300] env[61907]: DEBUG nova.compute.provider_tree [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.598070] env[61907]: WARNING nova.network.neutron [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] 9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b already exists in list: networks containing: ['9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b']. ignoring it [ 988.691176] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Successfully updated port: 1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 988.778831] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.812168] env[61907]: INFO nova.compute.manager [-] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Took 1.24 seconds to deallocate network for instance. [ 988.913099] env[61907]: DEBUG nova.scheduler.client.report [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 988.920803] env[61907]: DEBUG nova.network.neutron [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "address": "fa:16:3e:4a:e6:f4", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404d6d57-b8", "ovs_interfaceid": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.010385] env[61907]: DEBUG nova.compute.manager [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 989.010653] env[61907]: DEBUG nova.compute.manager [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing instance network info cache due to event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 989.010915] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.011123] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.011341] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.193599] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.193779] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.193939] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 989.280216] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.319827] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.418569] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.814s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.425258] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.105s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.425524] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.427459] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.428893] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.429118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.429635] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.429828] env[61907]: DEBUG nova.network.neutron [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Refreshing network info cache for port 404d6d57-b8d5-42c7-b255-d3b74e4f5797 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.431665] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-680b719d-a9da-4369-bee3-4b98c67fe530 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.451430] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 989.451921] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 989.452290] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 989.452673] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 989.453021] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 989.453340] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 989.453792] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 989.454157] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 989.454512] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 989.454885] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 989.455264] env[61907]: DEBUG nova.virt.hardware [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 989.462418] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfiguring VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 989.464422] env[61907]: INFO nova.scheduler.client.report [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Deleted allocations for instance 5475c612-e718-49eb-9760-9cfedbd7931d [ 989.466319] env[61907]: INFO nova.scheduler.client.report [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Deleted allocations for instance 6a092752-058f-4196-848e-58a94d084cc7 [ 989.467453] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f4255c5-7add-45f0-b580-eed547201d13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.496385] env[61907]: DEBUG oslo_vmware.api [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 989.496385] env[61907]: value = "task-1244247" [ 989.496385] env[61907]: _type = "Task" [ 989.496385] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.504462] env[61907]: DEBUG oslo_vmware.api [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244247, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.704167] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updated VIF entry in instance network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 989.704525] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.730656] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 989.789009] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.865410] env[61907]: DEBUG nova.network.neutron [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Updating instance_info_cache with network_info: [{"id": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "address": "fa:16:3e:39:7f:86", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9714e5-97", "ovs_interfaceid": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.993869] env[61907]: DEBUG oslo_concurrency.lockutils [None req-07a581a8-12c7-4b28-a696-f8fe747d19e7 tempest-ImagesTestJSON-1346768839 tempest-ImagesTestJSON-1346768839-project-member] Lock "6a092752-058f-4196-848e-58a94d084cc7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.728s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.994883] env[61907]: DEBUG oslo_concurrency.lockutils [None req-97f7c379-3671-4bf9-b5e9-e14b48808f47 tempest-DeleteServersTestJSON-97354578 tempest-DeleteServersTestJSON-97354578-project-member] Lock "5475c612-e718-49eb-9760-9cfedbd7931d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.068s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.004957] env[61907]: DEBUG oslo_vmware.api [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244247, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.170702] env[61907]: DEBUG nova.network.neutron [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updated VIF entry in instance network info cache for port 404d6d57-b8d5-42c7-b255-d3b74e4f5797. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.171277] env[61907]: DEBUG nova.network.neutron [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "address": "fa:16:3e:4a:e6:f4", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap404d6d57-b8", "ovs_interfaceid": "404d6d57-b8d5-42c7-b255-d3b74e4f5797", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.207501] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.208196] env[61907]: DEBUG nova.compute.manager [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 990.208196] env[61907]: DEBUG nova.compute.manager [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing instance network info cache due to event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 990.208196] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.208488] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.208488] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.271478] env[61907]: DEBUG nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Received event network-vif-deleted-e44f586e-da28-4e87-95ea-518bbf489218 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 990.271852] env[61907]: DEBUG nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Received event network-vif-plugged-1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 990.272100] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Acquiring lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.272318] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.272494] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.272704] env[61907]: DEBUG nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] No waiting events found dispatching network-vif-plugged-1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 990.272953] env[61907]: WARNING nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Received unexpected event network-vif-plugged-1d9714e5-97cf-4c24-a44e-6862c65cc023 for instance with vm_state building and task_state spawning. [ 990.273153] env[61907]: DEBUG nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Received event network-changed-1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 990.273318] env[61907]: DEBUG nova.compute.manager [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Refreshing instance network info cache due to event network-changed-1d9714e5-97cf-4c24-a44e-6862c65cc023. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 990.273489] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Acquiring lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.283348] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.283578] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.285919] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244244, 'name': CreateVM_Task, 'duration_secs': 3.233516} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.286100] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.286759] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.286952] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.287303] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.287629] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aaa85f5a-2cc7-4381-8789-84b24d0fdf80 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.293471] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 990.293471] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f16045-fa6b-883a-0a88-5dfd4ba02482" [ 990.293471] env[61907]: _type = "Task" [ 990.293471] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.302793] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f16045-fa6b-883a-0a88-5dfd4ba02482, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.367593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.367959] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Instance network_info: |[{"id": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "address": "fa:16:3e:39:7f:86", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9714e5-97", "ovs_interfaceid": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 990.368312] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Acquired lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.368508] env[61907]: DEBUG nova.network.neutron [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Refreshing network info cache for port 1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 990.370216] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:7f:86', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '16c6ea68-9b0e-4ac0-a484-7a9a40533017', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d9714e5-97cf-4c24-a44e-6862c65cc023', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 990.378720] env[61907]: DEBUG oslo.service.loopingcall [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 990.379224] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 990.379790] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f80afd01-6104-4cb4-8d88-e13ae7e7ce81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.400637] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 990.400637] env[61907]: value = "task-1244248" [ 990.400637] env[61907]: _type = "Task" [ 990.400637] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.410823] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244248, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.506304] env[61907]: DEBUG oslo_vmware.api [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244247, 'name': ReconfigVM_Task, 'duration_secs': 0.893105} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.506963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.507316] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfigured VM to attach interface {{(pid=61907) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 990.675105] env[61907]: DEBUG oslo_concurrency.lockutils [req-353ca4c3-5dad-48c5-8758-669f0b5a91eb req-47ca7206-e4c8-4768-a14e-76efd00d5cf8 service nova] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.786558] env[61907]: INFO nova.compute.manager [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Detaching volume f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2 [ 990.807790] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f16045-fa6b-883a-0a88-5dfd4ba02482, 'name': SearchDatastore_Task, 'duration_secs': 0.021932} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.808181] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.808488] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.808771] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.808959] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.809216] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.809520] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67db06dc-836a-46eb-be1b-a5ac1137d76e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.817422] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.817661] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.818434] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2de4ce39-7942-43fb-b267-cb4ad8543355 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.825661] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 990.825661] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ada7d2-56a9-772e-0261-d8a464d35a0a" [ 990.825661] env[61907]: _type = "Task" [ 990.825661] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.833154] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ada7d2-56a9-772e-0261-d8a464d35a0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.841098] env[61907]: INFO nova.virt.block_device [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Attempting to driver detach volume f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2 from mountpoint /dev/sdb [ 990.841326] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Volume detach. Driver type: vmdk {{(pid=61907) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 990.841516] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268293', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'name': 'volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '53176c9e-d15c-49d5-b4a9-22b780279ecb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'serial': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 990.842316] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d9e7ac-5579-4e4b-8fbf-2cb7c55325fe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.866047] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-787842e2-edbd-49cf-b225-9746e88b16ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.873671] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33b88079-061b-4cdd-82c8-6da61627f5c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.898428] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bada0e2-35f1-4aad-bfa1-cd83e6dd9d7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.918148] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] The volume has not been displaced from its original location: [datastore2] volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2/volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2.vmdk. No consolidation needed. {{(pid=61907) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 990.923262] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 990.926441] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbf72524-ca00-42e7-8209-ab7778496e33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.938665] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244248, 'name': CreateVM_Task, 'duration_secs': 0.404684} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.938825] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.939909] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.940097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.940423] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.940678] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04e92db1-d822-44e2-95c6-452621c40273 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.944039] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 990.944039] env[61907]: value = "task-1244249" [ 990.944039] env[61907]: _type = "Task" [ 990.944039] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.945297] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 990.945297] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ed9673-cfeb-f2e9-1793-373ba303eb73" [ 990.945297] env[61907]: _type = "Task" [ 990.945297] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.959824] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244249, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.963318] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ed9673-cfeb-f2e9-1793-373ba303eb73, 'name': SearchDatastore_Task, 'duration_secs': 0.009685} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.963625] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.963862] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.964112] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.964267] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.964449] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.964700] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b1d62036-917a-4ef8-a36d-a204cd6b07a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.972510] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.972636] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.973347] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2488dc5-51ab-450d-880c-cdf2161b440c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.979448] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 990.979448] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be034b-8c8a-1f13-65e6-f830ccb4c226" [ 990.979448] env[61907]: _type = "Task" [ 990.979448] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.987307] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be034b-8c8a-1f13-65e6-f830ccb4c226, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.013920] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a0ca2abe-1d04-4729-bfb9-ff9965efdaee tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.651s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.044826] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updated VIF entry in instance network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.045218] env[61907]: DEBUG nova.network.neutron [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.049605] env[61907]: DEBUG nova.compute.manager [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 991.049910] env[61907]: DEBUG nova.compute.manager [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing instance network info cache due to event network-changed-b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 991.050069] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] Acquiring lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.095775] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.096127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.096414] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.096796] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.097052] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.099155] env[61907]: INFO nova.compute.manager [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Terminating instance [ 991.166377] env[61907]: DEBUG nova.network.neutron [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Updated VIF entry in instance network info cache for port 1d9714e5-97cf-4c24-a44e-6862c65cc023. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 991.166807] env[61907]: DEBUG nova.network.neutron [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Updating instance_info_cache with network_info: [{"id": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "address": "fa:16:3e:39:7f:86", "network": {"id": "368f236c-5a0a-4d28-b378-f9a250afc983", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1981417647-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b57c1194e0f14ae498d99d302da346c5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "16c6ea68-9b0e-4ac0-a484-7a9a40533017", "external-id": "nsx-vlan-transportzone-384", "segmentation_id": 384, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d9714e5-97", "ovs_interfaceid": "1d9714e5-97cf-4c24-a44e-6862c65cc023", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.336061] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ada7d2-56a9-772e-0261-d8a464d35a0a, 'name': SearchDatastore_Task, 'duration_secs': 0.00957} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.336489] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91a32793-9c25-463d-b85f-f7ca79708b7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.341220] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 991.341220] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215272c-92e1-e03a-f048-88404274f597" [ 991.341220] env[61907]: _type = "Task" [ 991.341220] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.348456] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215272c-92e1-e03a-f048-88404274f597, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.454462] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244249, 'name': ReconfigVM_Task, 'duration_secs': 0.267229} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.454861] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 991.460042] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d10c45f0-a3b3-4f50-bd7e-82d3dcdf2bbd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.475024] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 991.475024] env[61907]: value = "task-1244250" [ 991.475024] env[61907]: _type = "Task" [ 991.475024] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.486248] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244250, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.491469] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52be034b-8c8a-1f13-65e6-f830ccb4c226, 'name': SearchDatastore_Task, 'duration_secs': 0.009728} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.492210] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-50bba1d5-517e-45b0-b261-3a50696f9ded {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.498270] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 991.498270] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d13347-fec5-cb0b-3e7d-e16e3176a4c3" [ 991.498270] env[61907]: _type = "Task" [ 991.498270] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.505324] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d13347-fec5-cb0b-3e7d-e16e3176a4c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.548222] env[61907]: DEBUG oslo_concurrency.lockutils [req-9d33c6bc-f5e6-4b68-a5a2-e6e2689eca6a req-5a9f4234-ae84-4411-aad4-12f0f31dc608 service nova] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.548756] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] Acquired lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.548992] env[61907]: DEBUG nova.network.neutron [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Refreshing network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.602623] env[61907]: DEBUG nova.compute.manager [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 991.602883] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 991.603788] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db7cb9a2-4292-4ab6-bd9e-dae6a6b264a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.612438] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 991.612690] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-304b40a7-f41e-4f8f-995f-c4ab22fa5904 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.618195] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 991.618195] env[61907]: value = "task-1244251" [ 991.618195] env[61907]: _type = "Task" [ 991.618195] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.626241] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.669997] env[61907]: DEBUG oslo_concurrency.lockutils [req-22bc766f-6da9-4ce4-8474-9510bd50e004 req-72172c86-0d06-4c6b-b9b9-b3b509086b7f service nova] Releasing lock "refresh_cache-696c8d63-5268-4239-8d41-079e9f3f7c66" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.851882] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5215272c-92e1-e03a-f048-88404274f597, 'name': SearchDatastore_Task, 'duration_secs': 0.025013} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.852167] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.852437] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.852731] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c4210692-4533-4151-95df-bf5d719de911 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.859876] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 991.859876] env[61907]: value = "task-1244252" [ 991.859876] env[61907]: _type = "Task" [ 991.859876] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.868278] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244252, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.983706] env[61907]: DEBUG oslo_vmware.api [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244250, 'name': ReconfigVM_Task, 'duration_secs': 0.137531} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.984064] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268293', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'name': 'volume-f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '53176c9e-d15c-49d5-b4a9-22b780279ecb', 'attached_at': '', 'detached_at': '', 'volume_id': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2', 'serial': 'f6f6196e-7d3e-4ec9-bf07-d6fb58c80cc2'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 992.007763] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d13347-fec5-cb0b-3e7d-e16e3176a4c3, 'name': SearchDatastore_Task, 'duration_secs': 0.012158} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.008064] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.008358] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 696c8d63-5268-4239-8d41-079e9f3f7c66/696c8d63-5268-4239-8d41-079e9f3f7c66.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 992.008650] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ba7e485-c835-4199-821d-689a1af50f92 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.017484] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 992.017484] env[61907]: value = "task-1244253" [ 992.017484] env[61907]: _type = "Task" [ 992.017484] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.023377] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.131883] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244251, 'name': PowerOffVM_Task, 'duration_secs': 0.387665} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.131883] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 992.131883] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 992.131883] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60545055-3a9b-44ad-9eb2-9b26f4a918d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.202819] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 992.205149] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 992.205149] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Deleting the datastore file [datastore2] b43efe48-95de-46ec-8cbb-c24cf7bd68a1 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 992.205149] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1232d35-42f9-4d0b-b285-58d2b5ebba0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.213247] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for the task: (returnval){ [ 992.213247] env[61907]: value = "task-1244255" [ 992.213247] env[61907]: _type = "Task" [ 992.213247] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.223741] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.372682] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244252, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.528050] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244253, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.552136] env[61907]: DEBUG nova.objects.instance [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'flavor' on Instance uuid 53176c9e-d15c-49d5-b4a9-22b780279ecb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.570587] env[61907]: DEBUG nova.network.neutron [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updated VIF entry in instance network info cache for port b79b943f-5924-4ec0-bf11-eda9d8da3285. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.571070] env[61907]: DEBUG nova.network.neutron [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [{"id": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "address": "fa:16:3e:9b:db:a6", "network": {"id": "76b6957b-e9ba-44d0-96eb-6ae2d5144528", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1810194610-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "a2fb5c5950d5400aa93060c09137c247", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e7a44713-0af1-486e-bc0d-00e03a769fa4", "external-id": "nsx-vlan-transportzone-420", "segmentation_id": 420, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb79b943f-59", "ovs_interfaceid": "b79b943f-5924-4ec0-bf11-eda9d8da3285", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.722499] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.736279] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.736748] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.871359] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244252, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515199} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.871998] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.872419] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.872848] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d20f499-62b2-4470-9dd9-a57103a2a2d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.883026] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 992.883026] env[61907]: value = "task-1244256" [ 992.883026] env[61907]: _type = "Task" [ 992.883026] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.892023] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244256, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.025843] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.678786} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.026479] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 696c8d63-5268-4239-8d41-079e9f3f7c66/696c8d63-5268-4239-8d41-079e9f3f7c66.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 993.029540] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 993.029540] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4831ab35-ab50-4547-9cdc-58da80ff73ac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.038029] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 993.038029] env[61907]: value = "task-1244257" [ 993.038029] env[61907]: _type = "Task" [ 993.038029] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.048539] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.074705] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a1a2f28-aa63-4a3f-b53d-7a9552b80d67 req-b7d36a0f-b066-42b0-9cf5-753b8a8e3207 service nova] Releasing lock "refresh_cache-b43efe48-95de-46ec-8cbb-c24cf7bd68a1" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.230724] env[61907]: DEBUG oslo_vmware.api [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Task: {'id': task-1244255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.875957} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.231026] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 993.231227] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 993.231404] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 993.231584] env[61907]: INFO nova.compute.manager [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Took 1.63 seconds to destroy the instance on the hypervisor. [ 993.231827] env[61907]: DEBUG oslo.service.loopingcall [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.232029] env[61907]: DEBUG nova.compute.manager [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 993.232124] env[61907]: DEBUG nova.network.neutron [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 993.241505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.241674] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.244589] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af27a9bb-fa38-41f6-a54a-4904d567ffa2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.261948] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b37179-ab90-4ad5-9256-98e66f5a59c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.288399] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfiguring VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 993.288804] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78551b25-5085-4e6e-9471-a4d9444de77d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.310060] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 993.310060] env[61907]: value = "task-1244258" [ 993.310060] env[61907]: _type = "Task" [ 993.310060] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.319710] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.396509] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244256, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088308} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.397015] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.398385] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6e0000c-7dcb-47d9-a731-b34c0544e679 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.426954] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.426954] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3419ec0-3dd2-4ed9-8c65-ee657309cd20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.448658] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 993.448658] env[61907]: value = "task-1244259" [ 993.448658] env[61907]: _type = "Task" [ 993.448658] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.459184] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.552683] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069056} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.552963] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.555139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b75f8c-8e0d-461a-8c42-bef938568d66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.559433] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b1985ca8-e555-4a96-b0a0-75659f78495a tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.276s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.585515] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] 696c8d63-5268-4239-8d41-079e9f3f7c66/696c8d63-5268-4239-8d41-079e9f3f7c66.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.586651] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf7de88f-9cde-4113-a86d-25e7bdc562a9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.610368] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 993.610368] env[61907]: value = "task-1244260" [ 993.610368] env[61907]: _type = "Task" [ 993.610368] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.620253] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244260, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.806173] env[61907]: DEBUG nova.compute.manager [req-b94e6969-0d96-4d8a-80b5-607faf21877a req-45ca1eeb-720f-4593-8298-b5ebff117f95 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Received event network-vif-deleted-b79b943f-5924-4ec0-bf11-eda9d8da3285 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 993.806313] env[61907]: INFO nova.compute.manager [req-b94e6969-0d96-4d8a-80b5-607faf21877a req-45ca1eeb-720f-4593-8298-b5ebff117f95 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Neutron deleted interface b79b943f-5924-4ec0-bf11-eda9d8da3285; detaching it from the instance and deleting it from the info cache [ 993.806656] env[61907]: DEBUG nova.network.neutron [req-b94e6969-0d96-4d8a-80b5-607faf21877a req-45ca1eeb-720f-4593-8298-b5ebff117f95 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.822669] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.846511] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.846799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.847027] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.847207] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.847488] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.852417] env[61907]: INFO nova.compute.manager [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Terminating instance [ 993.959048] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244259, 'name': ReconfigVM_Task, 'duration_secs': 0.345812} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.959369] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 4430db3b-0276-42b6-a67e-386a164ddc0e/4430db3b-0276-42b6-a67e-386a164ddc0e.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.959960] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ef6d925-60e9-4cbd-a4ff-495bf91c6ac0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.967162] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.967224] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.967414] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.967669] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.967795] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.969668] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 993.969668] env[61907]: value = "task-1244261" [ 993.969668] env[61907]: _type = "Task" [ 993.969668] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.969992] env[61907]: INFO nova.compute.manager [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Terminating instance [ 993.982628] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244261, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.121597] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244260, 'name': ReconfigVM_Task, 'duration_secs': 0.323084} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.121923] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Reconfigured VM instance instance-0000005d to attach disk [datastore2] 696c8d63-5268-4239-8d41-079e9f3f7c66/696c8d63-5268-4239-8d41-079e9f3f7c66.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.122572] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8bf43d0d-327a-4e9c-b8bf-fa115eee5c31 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.129161] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 994.129161] env[61907]: value = "task-1244262" [ 994.129161] env[61907]: _type = "Task" [ 994.129161] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.137113] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244262, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.278234] env[61907]: DEBUG nova.network.neutron [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.311355] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70fbeb1e-0781-45d9-beec-9d5f5e720d96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.322828] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.326729] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90402b37-f028-405b-9b19-1049a6f92908 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.355339] env[61907]: DEBUG nova.compute.manager [req-b94e6969-0d96-4d8a-80b5-607faf21877a req-45ca1eeb-720f-4593-8298-b5ebff117f95 service nova] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Detach interface failed, port_id=b79b943f-5924-4ec0-bf11-eda9d8da3285, reason: Instance b43efe48-95de-46ec-8cbb-c24cf7bd68a1 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 994.356251] env[61907]: DEBUG nova.compute.manager [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 994.356453] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.357603] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf1e363-1217-46e8-b215-295ed55033c1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.364229] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.364564] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-230cfff7-1de5-4a06-b962-e31faa5932c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.371366] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 994.371366] env[61907]: value = "task-1244263" [ 994.371366] env[61907]: _type = "Task" [ 994.371366] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.378982] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244263, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.476582] env[61907]: DEBUG nova.compute.manager [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 994.476824] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.481189] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d685dbf2-5600-487b-b696-e9b82f2a9340 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.483292] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244261, 'name': Rename_Task, 'duration_secs': 0.14448} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.483563] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.484136] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e4f0783d-568a-4342-953a-e87f7d37251a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.487983] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 994.488230] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c88ba0d-896c-4372-a176-b36a4693f867 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.490474] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 994.490474] env[61907]: value = "task-1244264" [ 994.490474] env[61907]: _type = "Task" [ 994.490474] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.495060] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 994.495060] env[61907]: value = "task-1244265" [ 994.495060] env[61907]: _type = "Task" [ 994.495060] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.498036] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244264, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.507492] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.641739] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244262, 'name': Rename_Task, 'duration_secs': 0.149504} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.642133] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.642321] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8ff19ceb-fa29-4d4a-84ed-a0a64bcd3413 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.649655] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 994.649655] env[61907]: value = "task-1244266" [ 994.649655] env[61907]: _type = "Task" [ 994.649655] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.657791] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244266, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.781591] env[61907]: INFO nova.compute.manager [-] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Took 1.55 seconds to deallocate network for instance. [ 994.826879] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.881792] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244263, 'name': PowerOffVM_Task, 'duration_secs': 0.199273} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.882192] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 994.882374] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.882821] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c406ff9a-b1c2-471e-8864-88fee715d39a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.954865] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 994.955182] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 994.955483] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleting the datastore file [datastore2] 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.955921] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94e97a3f-33f2-46c4-8210-ef26193f4614 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.964020] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 994.964020] env[61907]: value = "task-1244268" [ 994.964020] env[61907]: _type = "Task" [ 994.964020] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.973887] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244268, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.000524] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244264, 'name': PowerOnVM_Task} progress is 90%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.008609] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244265, 'name': PowerOffVM_Task, 'duration_secs': 0.164657} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.008892] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 995.009081] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 995.009342] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbbf4be1-ea07-494b-8301-bf0d47f3a5ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.131588] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.132036] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.132136] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleting the datastore file [datastore1] 53176c9e-d15c-49d5-b4a9-22b780279ecb {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.132454] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2571fa5c-bdec-438e-b9bb-64fa4c33f7fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.141993] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 995.141993] env[61907]: value = "task-1244270" [ 995.141993] env[61907]: _type = "Task" [ 995.141993] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.149805] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244270, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.157708] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244266, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.290112] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.290453] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.290720] env[61907]: DEBUG nova.objects.instance [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lazy-loading 'resources' on Instance uuid b43efe48-95de-46ec-8cbb-c24cf7bd68a1 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.324091] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.475352] env[61907]: DEBUG oslo_vmware.api [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244268, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.214932} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.475774] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.475954] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.476226] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.476433] env[61907]: INFO nova.compute.manager [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 995.476690] env[61907]: DEBUG oslo.service.loopingcall [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.477084] env[61907]: DEBUG nova.compute.manager [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 995.477084] env[61907]: DEBUG nova.network.neutron [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.499986] env[61907]: DEBUG oslo_vmware.api [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244264, 'name': PowerOnVM_Task, 'duration_secs': 0.544649} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.500314] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.500525] env[61907]: DEBUG nova.compute.manager [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.501348] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7473a3ea-d8a2-495a-b59f-16e8c6ecadc6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.651497] env[61907]: DEBUG oslo_vmware.api [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244270, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.346273} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.655898] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.656107] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.656293] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.656468] env[61907]: INFO nova.compute.manager [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Took 1.18 seconds to destroy the instance on the hypervisor. [ 995.656702] env[61907]: DEBUG oslo.service.loopingcall [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.656905] env[61907]: DEBUG nova.compute.manager [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 995.656997] env[61907]: DEBUG nova.network.neutron [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.664194] env[61907]: DEBUG oslo_vmware.api [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244266, 'name': PowerOnVM_Task, 'duration_secs': 0.511218} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.664521] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.664739] env[61907]: INFO nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Took 7.88 seconds to spawn the instance on the hypervisor. [ 995.664922] env[61907]: DEBUG nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 995.665713] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f5ed7e-d6e2-4330-9a38-891ca6d37f84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.830883] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.982254] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c360bf-3081-4e1e-ae6c-abef9d1e68fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.991718] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1931ae8d-7642-4192-b51d-60f16e9542a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.035040] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a8db74-09b2-4cf3-9f2a-345715361b3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.040191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.042951] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caa1439-3f97-4b26-bb2c-8b81f00260d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.058703] env[61907]: DEBUG nova.compute.provider_tree [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.183670] env[61907]: INFO nova.compute.manager [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Took 15.96 seconds to build instance. [ 996.328496] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.428337] env[61907]: DEBUG nova.compute.manager [req-d0ba91a4-76fb-4002-b519-66a792f78fb2 req-659913e9-7d54-4a6c-8845-acde8c2cb41c service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Received event network-vif-deleted-5290ad3b-915d-49f5-b5e8-57438dd6ca7b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 996.430225] env[61907]: INFO nova.compute.manager [req-d0ba91a4-76fb-4002-b519-66a792f78fb2 req-659913e9-7d54-4a6c-8845-acde8c2cb41c service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Neutron deleted interface 5290ad3b-915d-49f5-b5e8-57438dd6ca7b; detaching it from the instance and deleting it from the info cache [ 996.430225] env[61907]: DEBUG nova.network.neutron [req-d0ba91a4-76fb-4002-b519-66a792f78fb2 req-659913e9-7d54-4a6c-8845-acde8c2cb41c service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.585758] env[61907]: ERROR nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] [req-bec56946-8b41-4a08-b150-2dba26b2f2ae] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 91bca385-a423-4ca4-9da0-aeb4615e22d3. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-bec56946-8b41-4a08-b150-2dba26b2f2ae"}]} [ 996.607780] env[61907]: DEBUG nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 996.640807] env[61907]: DEBUG nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 996.641135] env[61907]: DEBUG nova.compute.provider_tree [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 147, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.657947] env[61907]: DEBUG nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 996.683746] env[61907]: DEBUG nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 996.686174] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3fda1325-ae06-4f52-ba1f-74ee151363bb tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.472s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.831350] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.881349] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad04ddde-3e60-4575-a32c-9f62cf195fd7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.888186] env[61907]: DEBUG nova.network.neutron [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.893740] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae2b119-7e2c-4319-a9eb-dab915c79754 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.929762] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e096f17-7bdc-48e7-bccf-4300c196a5d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.932271] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e1bee524-a449-412c-8195-5f1ff3a73094 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.944367] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5492dc75-179b-4f0e-a743-e5ab0de51c66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.951394] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e89d3499-fd4b-4321-96c6-8a6859b1957d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.970760] env[61907]: DEBUG nova.compute.provider_tree [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 996.993015] env[61907]: DEBUG nova.compute.manager [req-d0ba91a4-76fb-4002-b519-66a792f78fb2 req-659913e9-7d54-4a6c-8845-acde8c2cb41c service nova] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Detach interface failed, port_id=5290ad3b-915d-49f5-b5e8-57438dd6ca7b, reason: Instance 4c70ef23-fd26-4e90-b227-13aa050ff46d could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 997.131738] env[61907]: DEBUG nova.network.neutron [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.231297] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "4430db3b-0276-42b6-a67e-386a164ddc0e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.231961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.231961] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.232130] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.232174] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.234609] env[61907]: INFO nova.compute.manager [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Terminating instance [ 997.331763] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.394171] env[61907]: INFO nova.compute.manager [-] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Took 1.92 seconds to deallocate network for instance. [ 997.511510] env[61907]: DEBUG nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updated inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with generation 119 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:972}} [ 997.511929] env[61907]: DEBUG nova.compute.provider_tree [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 generation from 119 to 120 during operation: update_inventory {{(pid=61907) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 997.512182] env[61907]: DEBUG nova.compute.provider_tree [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 997.634859] env[61907]: INFO nova.compute.manager [-] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Took 1.98 seconds to deallocate network for instance. [ 997.738739] env[61907]: DEBUG nova.compute.manager [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 997.739142] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.741225] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05641144-ce21-40bb-bed1-c2a9e2902e48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.749759] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.750094] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcb34ca5-1840-4799-852d-ba38090f508a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.757856] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 997.757856] env[61907]: value = "task-1244271" [ 997.757856] env[61907]: _type = "Task" [ 997.757856] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.765616] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244271, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.822140] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "696c8d63-5268-4239-8d41-079e9f3f7c66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.822746] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.823085] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.823315] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.823504] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.825806] env[61907]: INFO nova.compute.manager [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Terminating instance [ 997.840370] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.901986] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.017649] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.727s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.020109] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.981s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.020338] env[61907]: DEBUG nova.objects.instance [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 998.036394] env[61907]: INFO nova.scheduler.client.report [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Deleted allocations for instance b43efe48-95de-46ec-8cbb-c24cf7bd68a1 [ 998.142277] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.183588] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "6046a82d-f73d-4369-8208-c5a2e5c77431" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.183885] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.268760] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244271, 'name': PowerOffVM_Task, 'duration_secs': 0.217763} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.269064] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.269230] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.269486] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2f3a7ec9-1843-4d90-bb12-d32499820019 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.335928] env[61907]: DEBUG nova.compute.manager [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 998.336312] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 998.336669] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.337462] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18e584c8-fc9e-4bab-86ff-c39f1f4875c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.344793] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 998.345906] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-48cc03dc-cbc4-40c4-9b17-592090dfed46 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.347394] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.347645] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.347773] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleting the datastore file [datastore1] 4430db3b-0276-42b6-a67e-386a164ddc0e {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.347994] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3be48286-2975-4e03-aa31-e9ee0e5c70fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.353504] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 998.353504] env[61907]: value = "task-1244273" [ 998.353504] env[61907]: _type = "Task" [ 998.353504] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.357281] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 998.357281] env[61907]: value = "task-1244274" [ 998.357281] env[61907]: _type = "Task" [ 998.357281] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.363111] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244273, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.367637] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.461999] env[61907]: DEBUG nova.compute.manager [req-324857df-3879-4d80-a0cf-a2874a0be299 req-964dcec9-1f45-4725-80b9-728bccc39d3a service nova] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Received event network-vif-deleted-8c5ce2c9-4bf4-4aa0-9e00-23bfa89f4a53 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 998.546373] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ae3a5cd5-99f0-411e-8a3e-26da2e765cad tempest-ServerRescueTestJSONUnderV235-1475668192 tempest-ServerRescueTestJSONUnderV235-1475668192-project-member] Lock "b43efe48-95de-46ec-8cbb-c24cf7bd68a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.449s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.686986] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 998.836971] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.867861] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244273, 'name': PowerOffVM_Task, 'duration_secs': 0.173872} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.870632] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 998.870632] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 998.870880] env[61907]: DEBUG oslo_vmware.api [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197208} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.871395] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f42a62d0-952b-482c-af6c-cd725e366aa0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.874907] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.874907] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.874907] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.874907] env[61907]: INFO nova.compute.manager [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 998.874907] env[61907]: DEBUG oslo.service.loopingcall [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.874907] env[61907]: DEBUG nova.compute.manager [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 998.874907] env[61907]: DEBUG nova.network.neutron [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.954940] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 998.958156] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 998.958156] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleting the datastore file [datastore2] 696c8d63-5268-4239-8d41-079e9f3f7c66 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 998.958156] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-144a88ea-5be1-4a72-af45-9f7a8eaead96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.963248] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for the task: (returnval){ [ 998.963248] env[61907]: value = "task-1244276" [ 998.963248] env[61907]: _type = "Task" [ 998.963248] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.972525] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244276, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.030851] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f968633-3d6b-48b4-b2e8-b76086b4caba tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.034045] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.130s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.034045] env[61907]: DEBUG nova.objects.instance [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'resources' on Instance uuid 4c70ef23-fd26-4e90-b227-13aa050ff46d {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.211296] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.339707] env[61907]: DEBUG oslo_vmware.api [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244258, 'name': ReconfigVM_Task, 'duration_secs': 5.736546} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.340017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.340346] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Reconfigured VM to detach interface {{(pid=61907) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 999.473805] env[61907]: DEBUG oslo_vmware.api [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Task: {'id': task-1244276, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155213} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.474047] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 999.475156] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 999.475156] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 999.475328] env[61907]: INFO nova.compute.manager [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Took 1.14 seconds to destroy the instance on the hypervisor. [ 999.475573] env[61907]: DEBUG oslo.service.loopingcall [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 999.475790] env[61907]: DEBUG nova.compute.manager [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 999.475893] env[61907]: DEBUG nova.network.neutron [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 999.618023] env[61907]: DEBUG nova.network.neutron [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.710867] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f59c5cc-2805-4269-ab67-9fdcd7eb0b34 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.723017] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d66bb4-26a3-4cc1-b569-11de7f6c1d36 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.755265] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-242005bf-d89c-4a65-bcbc-3bfb79d80a50 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.764120] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ad06bb-c143-40da-bca7-c7d31dae0b71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.780787] env[61907]: DEBUG nova.compute.provider_tree [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.121503] env[61907]: INFO nova.compute.manager [-] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Took 1.25 seconds to deallocate network for instance. [ 1000.283980] env[61907]: DEBUG nova.scheduler.client.report [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1000.335684] env[61907]: DEBUG nova.network.neutron [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.496154] env[61907]: DEBUG nova.compute.manager [req-4d78791e-4745-4d26-91bb-2263dccd04aa req-d50ad09d-d581-479c-a4f6-d77f0c76f828 service nova] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Received event network-vif-deleted-bd27cf86-5b58-48e0-b10f-ae391a48f9f2 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1000.496703] env[61907]: DEBUG nova.compute.manager [req-4d78791e-4745-4d26-91bb-2263dccd04aa req-d50ad09d-d581-479c-a4f6-d77f0c76f828 service nova] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Received event network-vif-deleted-1d9714e5-97cf-4c24-a44e-6862c65cc023 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1000.627984] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.792960] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.761s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.795913] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.654s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.796217] env[61907]: DEBUG nova.objects.instance [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'resources' on Instance uuid 53176c9e-d15c-49d5-b4a9-22b780279ecb {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1000.806205] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1000.806492] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquired lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1000.806638] env[61907]: DEBUG nova.network.neutron [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1000.819996] env[61907]: INFO nova.scheduler.client.report [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleted allocations for instance 4c70ef23-fd26-4e90-b227-13aa050ff46d [ 1000.838615] env[61907]: INFO nova.compute.manager [-] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Took 1.36 seconds to deallocate network for instance. [ 1001.313593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.313951] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.314190] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.314384] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.314556] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.317398] env[61907]: INFO nova.compute.manager [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Terminating instance [ 1001.327138] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ffb5a170-4019-4593-afdd-099437d3c9c9 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "4c70ef23-fd26-4e90-b227-13aa050ff46d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.480s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.347688] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.514322] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ec4cd2-80af-40e8-9e90-40e811eeca13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.522199] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d43f1640-7775-4ea7-97d0-113001a08d7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.558141] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c681a780-c248-482c-b6c3-5f5e52cc5e3e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.565952] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae381ba6-aa20-4b94-ac84-ef187f8c0dc3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.579596] env[61907]: DEBUG nova.compute.provider_tree [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.630741] env[61907]: INFO nova.network.neutron [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Port 404d6d57-b8d5-42c7-b255-d3b74e4f5797 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1001.631143] env[61907]: DEBUG nova.network.neutron [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [{"id": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "address": "fa:16:3e:50:63:d2", "network": {"id": "9b58f94c-a9c4-453c-8ce1-f21d8a62ea5b", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-414120460-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6df6b1cd82e24d2f8aa1812575f03f8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0662d339-0c", "ovs_interfaceid": "0662d339-0c3c-49e2-9c37-6dd66f7e704e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1001.821905] env[61907]: DEBUG nova.compute.manager [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1001.822162] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1001.823084] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c9dea6-3d4d-4a0f-949b-7e924fde9233 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.831176] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1001.831424] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7ebc798-3da5-4cf2-b834-1e13ab7582d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.837685] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 1001.837685] env[61907]: value = "task-1244277" [ 1001.837685] env[61907]: _type = "Task" [ 1001.837685] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.845531] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244277, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.082643] env[61907]: DEBUG nova.scheduler.client.report [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1002.134459] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Releasing lock "refresh_cache-f50bf687-609c-44e1-bd75-cf4efaeadb71" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.348196] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244277, 'name': PowerOffVM_Task, 'duration_secs': 0.181693} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.348514] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1002.348693] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1002.348957] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-baddf39a-66a3-4b64-98c4-ed236f95a8be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.413579] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1002.413579] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1002.413579] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleting the datastore file [datastore1] f50bf687-609c-44e1-bd75-cf4efaeadb71 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1002.413579] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8511bb7c-65fc-40ed-ab74-2e7a341f1048 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.417922] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 1002.417922] env[61907]: value = "task-1244279" [ 1002.417922] env[61907]: _type = "Task" [ 1002.417922] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.425554] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244279, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.590075] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.792s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.590984] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.380s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.592655] env[61907]: INFO nova.compute.claims [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1002.610728] env[61907]: INFO nova.scheduler.client.report [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted allocations for instance 53176c9e-d15c-49d5-b4a9-22b780279ecb [ 1002.638400] env[61907]: DEBUG oslo_concurrency.lockutils [None req-2c78d4e1-d563-4666-9249-491c92848130 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "interface-f50bf687-609c-44e1-bd75-cf4efaeadb71-404d6d57-b8d5-42c7-b255-d3b74e4f5797" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.901s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.929830] env[61907]: DEBUG oslo_vmware.api [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244279, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139683} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.930184] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1002.930403] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1002.930601] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1002.930761] env[61907]: INFO nova.compute.manager [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1002.931022] env[61907]: DEBUG oslo.service.loopingcall [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.931238] env[61907]: DEBUG nova.compute.manager [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1002.931332] env[61907]: DEBUG nova.network.neutron [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1003.120444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-e3843ffd-95c8-41e9-8090-c0d9b15ddaef tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "53176c9e-d15c-49d5-b4a9-22b780279ecb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.153s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.657639] env[61907]: DEBUG nova.compute.manager [req-d598c8f9-db50-4bd8-9980-058a1040a1a2 req-6a868a29-a663-44af-a764-91564b57bbc5 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Received event network-vif-deleted-0662d339-0c3c-49e2-9c37-6dd66f7e704e {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1003.657639] env[61907]: INFO nova.compute.manager [req-d598c8f9-db50-4bd8-9980-058a1040a1a2 req-6a868a29-a663-44af-a764-91564b57bbc5 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Neutron deleted interface 0662d339-0c3c-49e2-9c37-6dd66f7e704e; detaching it from the instance and deleting it from the info cache [ 1003.657639] env[61907]: DEBUG nova.network.neutron [req-d598c8f9-db50-4bd8-9980-058a1040a1a2 req-6a868a29-a663-44af-a764-91564b57bbc5 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.699437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.700076] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.760239] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1006678c-6063-4864-a391-d151a50cd717 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.768014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81260f64-321e-4952-94a2-9fbc4bb93cef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.801548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac2e65bc-ee63-4d7f-a7ed-c69a46ef70af {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.808345] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-262661a1-006f-43a8-89b6-8421e177eb42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.823303] env[61907]: DEBUG nova.compute.provider_tree [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.130717] env[61907]: DEBUG nova.network.neutron [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.160483] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d96a77a7-a54c-4de4-b998-ce11c728053b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.170734] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-663e158b-bd01-48d9-b000-590d4984bbeb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.197281] env[61907]: DEBUG nova.compute.manager [req-d598c8f9-db50-4bd8-9980-058a1040a1a2 req-6a868a29-a663-44af-a764-91564b57bbc5 service nova] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Detach interface failed, port_id=0662d339-0c3c-49e2-9c37-6dd66f7e704e, reason: Instance f50bf687-609c-44e1-bd75-cf4efaeadb71 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1004.204894] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1004.326197] env[61907]: DEBUG nova.scheduler.client.report [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1004.633411] env[61907]: INFO nova.compute.manager [-] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Took 1.70 seconds to deallocate network for instance. [ 1004.726593] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.831903] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.241s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.832472] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1004.835207] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.207s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.835461] env[61907]: DEBUG nova.objects.instance [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'resources' on Instance uuid 4430db3b-0276-42b6-a67e-386a164ddc0e {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.140114] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.341527] env[61907]: DEBUG nova.compute.utils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.343105] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1005.343288] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1005.389479] env[61907]: DEBUG nova.policy [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bddec4540fd14ad9bd763808c4eadbab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06abbb40481b4095adb361c642fa82b9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.459466] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e026b955-f5b6-4df1-b9a0-1878034e09ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.467277] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7c47dad-2708-412e-89fd-c309d1dbbb91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.498647] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d37acc-57f5-475f-8df1-c9ff923b1404 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.506218] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15397336-b526-49c4-a316-67c4c850e1ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.519449] env[61907]: DEBUG nova.compute.provider_tree [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1005.642993] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Successfully created port: 5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1005.846655] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1006.024785] env[61907]: DEBUG nova.scheduler.client.report [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1006.530238] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.695s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.532520] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.185s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.532767] env[61907]: DEBUG nova.objects.instance [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lazy-loading 'resources' on Instance uuid 696c8d63-5268-4239-8d41-079e9f3f7c66 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.554648] env[61907]: INFO nova.scheduler.client.report [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted allocations for instance 4430db3b-0276-42b6-a67e-386a164ddc0e [ 1006.857056] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1006.882627] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1006.882893] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1006.883069] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1006.883327] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1006.883494] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1006.883649] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1006.883862] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1006.884044] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1006.884229] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1006.884401] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1006.884582] env[61907]: DEBUG nova.virt.hardware [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1006.885473] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83561c02-cce8-4a55-8958-ca845fadb54a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.893742] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89b511cc-2d82-4195-938b-884fd6e4c127 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.014236] env[61907]: DEBUG nova.compute.manager [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Received event network-vif-plugged-5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1007.014450] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] Acquiring lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.014742] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.014888] env[61907]: DEBUG oslo_concurrency.lockutils [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.015086] env[61907]: DEBUG nova.compute.manager [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] No waiting events found dispatching network-vif-plugged-5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1007.015258] env[61907]: WARNING nova.compute.manager [req-ad7e4733-2fd6-499a-803b-86fe27c1f5ce req-623e7b9c-42a9-4d0f-b153-1839064fdcc8 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Received unexpected event network-vif-plugged-5e64c307-2d27-4555-9376-fd3d08be3fc0 for instance with vm_state building and task_state spawning. [ 1007.062905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-12153cf6-bf1e-42f8-9bad-22b246762257 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "4430db3b-0276-42b6-a67e-386a164ddc0e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.831s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.121618] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Successfully updated port: 5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1007.188060] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876d825e-7513-4df4-a535-72505fe71045 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.196859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a9c8da-f260-471d-9757-34521c3b1efc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.226691] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575c92bb-daab-4898-afbb-78676f3d1b9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.234585] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6775b2-a0f8-4fda-8b6a-e2e3feb9b0ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.247611] env[61907]: DEBUG nova.compute.provider_tree [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1007.573769] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "23526967-cd8f-4581-b9c5-1c270d385163" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.574154] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.574385] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "23526967-cd8f-4581-b9c5-1c270d385163-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.574577] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.574753] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.576964] env[61907]: INFO nova.compute.manager [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Terminating instance [ 1007.627625] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.627772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquired lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.627930] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1007.753292] env[61907]: DEBUG nova.scheduler.client.report [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1008.081090] env[61907]: DEBUG nova.compute.manager [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1008.081467] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1008.082318] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d28f70ba-8c8f-419f-ab93-44e2fa0d0f23 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.090414] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.090674] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eece317a-95b9-4059-9ff8-29260418d6cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.097822] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 1008.097822] env[61907]: value = "task-1244281" [ 1008.097822] env[61907]: _type = "Task" [ 1008.097822] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.106291] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244281, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.160502] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1008.256867] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.724s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.259250] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.533s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.260816] env[61907]: INFO nova.compute.claims [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1008.285725] env[61907]: INFO nova.scheduler.client.report [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Deleted allocations for instance 696c8d63-5268-4239-8d41-079e9f3f7c66 [ 1008.369936] env[61907]: DEBUG nova.network.neutron [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Updating instance_info_cache with network_info: [{"id": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "address": "fa:16:3e:5c:5b:76", "network": {"id": "138137c8-ccc9-41be-8296-badbfd931ea2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-931122603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06abbb40481b4095adb361c642fa82b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e64c307-2d", "ovs_interfaceid": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.611262] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244281, 'name': PowerOffVM_Task, 'duration_secs': 0.21321} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.611871] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1008.612177] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1008.612600] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9d2104dd-2169-4fe3-b53d-cc63071f582a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.700903] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1008.701123] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1008.701325] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleting the datastore file [datastore1] 23526967-cd8f-4581-b9c5-1c270d385163 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.701606] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10b346fa-541a-4c4d-899a-50ba8f087410 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.708936] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 1008.708936] env[61907]: value = "task-1244283" [ 1008.708936] env[61907]: _type = "Task" [ 1008.708936] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.716474] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244283, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.797132] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3f268cda-683f-45e7-8b66-9303c9b64561 tempest-ServerDiskConfigTestJSON-1300436059 tempest-ServerDiskConfigTestJSON-1300436059-project-member] Lock "696c8d63-5268-4239-8d41-079e9f3f7c66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.975s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.873783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Releasing lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.874159] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Instance network_info: |[{"id": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "address": "fa:16:3e:5c:5b:76", "network": {"id": "138137c8-ccc9-41be-8296-badbfd931ea2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-931122603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06abbb40481b4095adb361c642fa82b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e64c307-2d", "ovs_interfaceid": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1008.874621] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:5b:76', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd22cb4ec-277f-41ee-8aba-b3d54442b93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5e64c307-2d27-4555-9376-fd3d08be3fc0', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1008.882136] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Creating folder: Project (06abbb40481b4095adb361c642fa82b9). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.882737] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3a31b455-a43d-4f81-8374-4c746edf6ec3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.893759] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Created folder: Project (06abbb40481b4095adb361c642fa82b9) in parent group-v268168. [ 1008.893941] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Creating folder: Instances. Parent ref: group-v268309. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1008.894193] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4f5140f4-07a8-4664-87e2-90ab2c772f9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.902315] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Created folder: Instances in parent group-v268309. [ 1008.902521] env[61907]: DEBUG oslo.service.loopingcall [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.902693] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1008.902903] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8af63ca0-bc22-468c-9f5a-b22776bed1df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.922594] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1008.922594] env[61907]: value = "task-1244286" [ 1008.922594] env[61907]: _type = "Task" [ 1008.922594] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.936254] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244286, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.041745] env[61907]: DEBUG nova.compute.manager [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Received event network-changed-5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1009.041912] env[61907]: DEBUG nova.compute.manager [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Refreshing instance network info cache due to event network-changed-5e64c307-2d27-4555-9376-fd3d08be3fc0. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1009.042277] env[61907]: DEBUG oslo_concurrency.lockutils [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] Acquiring lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.042436] env[61907]: DEBUG oslo_concurrency.lockutils [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] Acquired lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.042609] env[61907]: DEBUG nova.network.neutron [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Refreshing network info cache for port 5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1009.220176] env[61907]: DEBUG oslo_vmware.api [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244283, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.198503} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.220676] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.220676] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1009.220903] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1009.221123] env[61907]: INFO nova.compute.manager [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1009.221385] env[61907]: DEBUG oslo.service.loopingcall [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.221582] env[61907]: DEBUG nova.compute.manager [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1009.221678] env[61907]: DEBUG nova.network.neutron [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1009.394943] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c1f5389-b40f-49dd-8ad0-64a08e6fac74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.403313] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09546411-2704-4712-b6ee-a54d86df3db8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.437725] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785e1874-4b49-44b5-9bc8-29c6dcf496c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.448760] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de5a7e4-ce71-453a-8efe-b7f26d140eff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.453090] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244286, 'name': CreateVM_Task, 'duration_secs': 0.338989} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.453192] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1009.454380] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.454555] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.454882] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1009.455281] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bea32022-54ef-4d3b-91af-ee1563dfb32b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.466388] env[61907]: DEBUG nova.compute.provider_tree [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.469017] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1009.469017] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52853100-f7a2-9316-0be2-51c8200f8a5c" [ 1009.469017] env[61907]: _type = "Task" [ 1009.469017] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.479669] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52853100-f7a2-9316-0be2-51c8200f8a5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.495968] env[61907]: DEBUG nova.compute.manager [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1009.910140] env[61907]: DEBUG nova.network.neutron [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Updated VIF entry in instance network info cache for port 5e64c307-2d27-4555-9376-fd3d08be3fc0. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1009.910529] env[61907]: DEBUG nova.network.neutron [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Updating instance_info_cache with network_info: [{"id": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "address": "fa:16:3e:5c:5b:76", "network": {"id": "138137c8-ccc9-41be-8296-badbfd931ea2", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-931122603-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06abbb40481b4095adb361c642fa82b9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d22cb4ec-277f-41ee-8aba-b3d54442b93d", "external-id": "nsx-vlan-transportzone-652", "segmentation_id": 652, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5e64c307-2d", "ovs_interfaceid": "5e64c307-2d27-4555-9376-fd3d08be3fc0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.967206] env[61907]: DEBUG nova.scheduler.client.report [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1009.980064] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52853100-f7a2-9316-0be2-51c8200f8a5c, 'name': SearchDatastore_Task, 'duration_secs': 0.010208} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.980387] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.980624] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1009.980867] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1009.981032] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1009.981231] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.982026] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0aac8858-c1f5-49bb-a84a-25108cf5d198 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.991023] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.991023] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1009.991649] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18c57694-986a-41a5-9ccf-0583fb875a0b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.997132] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1009.997132] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528c3b0b-5cef-355f-9ca5-8ec806c6d046" [ 1009.997132] env[61907]: _type = "Task" [ 1009.997132] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.009961] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528c3b0b-5cef-355f-9ca5-8ec806c6d046, 'name': SearchDatastore_Task, 'duration_secs': 0.008949} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.010699] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eedc59b2-b88d-4554-88b1-38db63e77fbf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.015712] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1010.015712] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526e58e4-cacc-f763-e00d-2c514d3ec304" [ 1010.015712] env[61907]: _type = "Task" [ 1010.015712] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.021106] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.025464] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526e58e4-cacc-f763-e00d-2c514d3ec304, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.175775] env[61907]: DEBUG nova.network.neutron [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.413114] env[61907]: DEBUG oslo_concurrency.lockutils [req-071dbe37-257f-4f0b-8466-5d8aa3dcf6bb req-1dc1a629-247a-457c-a886-3e3946e61951 service nova] Releasing lock "refresh_cache-6046a82d-f73d-4369-8208-c5a2e5c77431" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.474846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.216s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.475464] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1010.478424] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.338s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.478663] env[61907]: DEBUG nova.objects.instance [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'resources' on Instance uuid f50bf687-609c-44e1-bd75-cf4efaeadb71 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.528154] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526e58e4-cacc-f763-e00d-2c514d3ec304, 'name': SearchDatastore_Task, 'duration_secs': 0.009426} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.528546] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.529066] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6046a82d-f73d-4369-8208-c5a2e5c77431/6046a82d-f73d-4369-8208-c5a2e5c77431.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1010.529171] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-248c1d07-c9d2-465d-8940-9816bd4e68c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.537168] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1010.537168] env[61907]: value = "task-1244287" [ 1010.537168] env[61907]: _type = "Task" [ 1010.537168] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.545387] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244287, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.678696] env[61907]: INFO nova.compute.manager [-] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Took 1.46 seconds to deallocate network for instance. [ 1010.982119] env[61907]: DEBUG nova.compute.utils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.986073] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1010.986264] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.035507] env[61907]: DEBUG nova.policy [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '575a48c6fb4d44638edf42db3763e78a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a85baf8058442b39a07436a046e1ebe', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.045904] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244287, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.470934} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.048238] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6046a82d-f73d-4369-8208-c5a2e5c77431/6046a82d-f73d-4369-8208-c5a2e5c77431.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1011.048465] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1011.048890] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a695f9f-cf66-4784-98bf-9007e6b0c928 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.055144] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1011.055144] env[61907]: value = "task-1244288" [ 1011.055144] env[61907]: _type = "Task" [ 1011.055144] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.067690] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244288, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.069948] env[61907]: DEBUG nova.compute.manager [req-060b16ed-5cbb-427f-af96-740ca474b38e req-93d2fab9-b4e7-4b42-bbcc-9c75f06f52fd service nova] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Received event network-vif-deleted-1773082b-e928-4ca3-ab1c-515606cb7e12 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1011.118276] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b8f413-8ae6-43db-8d6d-3f62991f023a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.129209] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b5109c-4bff-47a4-8ea5-87b7fb1a06ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.159622] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66836bca-f42a-4cee-82c3-79f276f46cb8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.169240] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32e5fbd4-47f5-45b1-a2b2-d250a04ebf45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.182973] env[61907]: DEBUG nova.compute.provider_tree [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.187536] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.318921] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Successfully created port: cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1011.489620] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1011.565020] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244288, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071217} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.565325] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1011.566123] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10baa3d7-c965-4920-bf5c-855bcf6093c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.587626] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 6046a82d-f73d-4369-8208-c5a2e5c77431/6046a82d-f73d-4369-8208-c5a2e5c77431.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1011.587878] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dc4c241d-30e5-42a1-9dca-a0ca9a8943d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.606456] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1011.606456] env[61907]: value = "task-1244289" [ 1011.606456] env[61907]: _type = "Task" [ 1011.606456] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.616334] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244289, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.686230] env[61907]: DEBUG nova.scheduler.client.report [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1012.116343] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244289, 'name': ReconfigVM_Task, 'duration_secs': 0.295013} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.116637] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 6046a82d-f73d-4369-8208-c5a2e5c77431/6046a82d-f73d-4369-8208-c5a2e5c77431.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1012.117316] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a937a53-ec81-4ea5-9f31-751abc073d99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.123473] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1012.123473] env[61907]: value = "task-1244290" [ 1012.123473] env[61907]: _type = "Task" [ 1012.123473] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.131057] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244290, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.191592] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.193975] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.174s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.220066] env[61907]: INFO nova.scheduler.client.report [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted allocations for instance f50bf687-609c-44e1-bd75-cf4efaeadb71 [ 1012.498922] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1012.523527] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1012.523789] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1012.523954] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1012.524163] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1012.524320] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1012.524473] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1012.524687] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1012.524858] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1012.525048] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1012.525230] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1012.525413] env[61907]: DEBUG nova.virt.hardware [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1012.526336] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176c75a4-c313-4e84-a4c9-a2f6fd4426b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.534051] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983d17c2-4b98-48b6-8cb7-53a61c2d39b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.633783] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244290, 'name': Rename_Task, 'duration_secs': 0.148597} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.634082] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1012.634338] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbcd540c-e9b7-4700-82cd-fc0e72495c12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.640152] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1012.640152] env[61907]: value = "task-1244291" [ 1012.640152] env[61907]: _type = "Task" [ 1012.640152] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.649056] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244291, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.699109] env[61907]: INFO nova.compute.claims [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.727992] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7590e080-5e5c-4289-913c-c74df6aa1fd2 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "f50bf687-609c-44e1-bd75-cf4efaeadb71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.414s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.793557] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Successfully updated port: cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.093665] env[61907]: DEBUG nova.compute.manager [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Received event network-vif-plugged-cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1013.093926] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.094155] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.094349] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.094530] env[61907]: DEBUG nova.compute.manager [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] No waiting events found dispatching network-vif-plugged-cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1013.094740] env[61907]: WARNING nova.compute.manager [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Received unexpected event network-vif-plugged-cd2253b4-0244-45b6-80cb-b483ff82a146 for instance with vm_state building and task_state spawning. [ 1013.094917] env[61907]: DEBUG nova.compute.manager [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Received event network-changed-cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1013.095092] env[61907]: DEBUG nova.compute.manager [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Refreshing instance network info cache due to event network-changed-cd2253b4-0244-45b6-80cb-b483ff82a146. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1013.095367] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Acquiring lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.095555] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Acquired lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.095733] env[61907]: DEBUG nova.network.neutron [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Refreshing network info cache for port cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1013.150829] env[61907]: DEBUG oslo_vmware.api [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244291, 'name': PowerOnVM_Task, 'duration_secs': 0.453681} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.151294] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1013.151578] env[61907]: INFO nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Took 6.29 seconds to spawn the instance on the hypervisor. [ 1013.151801] env[61907]: DEBUG nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1013.152582] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0f7efc-65a5-4634-9d2d-1e8b72b20140 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.205571] env[61907]: INFO nova.compute.resource_tracker [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating resource usage from migration d4310bd1-9dd8-406e-a220-f078e7476274 [ 1013.296488] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.305443] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241382b0-e47c-4cb0-9993-760ddd7de252 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.313377] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-117d9983-0c3d-4dd8-ab03-5e9c6e68f7c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.343401] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5adb13-6ced-4de3-86d6-129ff10def07 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.350699] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21484022-a3bb-4c9b-9ee9-b72e4c760ecb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.363706] env[61907]: DEBUG nova.compute.provider_tree [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.485090] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.485394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.485623] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "fcde07bf-858a-4377-b27c-4f17356306af-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.485814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.485994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.488186] env[61907]: INFO nova.compute.manager [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Terminating instance [ 1013.637974] env[61907]: DEBUG nova.network.neutron [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.670440] env[61907]: INFO nova.compute.manager [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Took 14.48 seconds to build instance. [ 1013.715138] env[61907]: DEBUG nova.network.neutron [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.866694] env[61907]: DEBUG nova.scheduler.client.report [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1013.991774] env[61907]: DEBUG nova.compute.manager [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1013.992088] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1013.993011] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45950e92-1415-4f21-8103-3ff257e77dc1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.003030] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.003156] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95be4649-30c9-45c0-a8c8-27236a58b249 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.009700] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 1014.009700] env[61907]: value = "task-1244292" [ 1014.009700] env[61907]: _type = "Task" [ 1014.009700] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.017087] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.172182] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8194e101-3b19-43ef-b20d-b49ff6f0a58c tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.988s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.218070] env[61907]: DEBUG oslo_concurrency.lockutils [req-114087dd-cec3-4dd5-aa14-83cf28c15168 req-3eb54590-f895-4f30-8c13-e6fdda46f883 service nova] Releasing lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.218574] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.218818] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.371854] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.178s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.372112] env[61907]: INFO nova.compute.manager [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Migrating [ 1014.378645] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.191s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.378881] env[61907]: DEBUG nova.objects.instance [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'resources' on Instance uuid 23526967-cd8f-4581-b9c5-1c270d385163 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.520057] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244292, 'name': PowerOffVM_Task, 'duration_secs': 0.207696} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.520306] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.520480] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1014.520734] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f42ed1b0-5522-49c2-b569-2c82df18196d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.578740] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1014.578962] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1014.579168] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleting the datastore file [datastore1] fcde07bf-858a-4377-b27c-4f17356306af {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1014.579441] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3639626b-c3a5-4cc1-99d6-05a2e79e96ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.586088] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for the task: (returnval){ [ 1014.586088] env[61907]: value = "task-1244294" [ 1014.586088] env[61907]: _type = "Task" [ 1014.586088] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.593648] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244294, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.749897] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.768191] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "6046a82d-f73d-4369-8208-c5a2e5c77431" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.768601] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.768836] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.769046] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.769228] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.772856] env[61907]: INFO nova.compute.manager [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Terminating instance [ 1014.888745] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.888938] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.889135] env[61907]: DEBUG nova.network.neutron [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.992656] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79126703-32d6-4e44-93c8-da92480b136c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.001052] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52f66bd-fa7e-4e26-8790-930ac8a4c6f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.035142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a695bc-4b28-4c63-912b-4ca0d113145a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.042552] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d787114-d8f2-43ba-981b-fd02675b4e88 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.055289] env[61907]: DEBUG nova.compute.provider_tree [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.094749] env[61907]: DEBUG oslo_vmware.api [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Task: {'id': task-1244294, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152414} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.095185] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1015.095279] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1015.095440] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1015.095605] env[61907]: INFO nova.compute.manager [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1015.095851] env[61907]: DEBUG oslo.service.loopingcall [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.096419] env[61907]: DEBUG nova.compute.manager [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1015.096520] env[61907]: DEBUG nova.network.neutron [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1015.181823] env[61907]: DEBUG nova.network.neutron [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating instance_info_cache with network_info: [{"id": "cd2253b4-0244-45b6-80cb-b483ff82a146", "address": "fa:16:3e:35:ff:d4", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2253b4-02", "ovs_interfaceid": "cd2253b4-0244-45b6-80cb-b483ff82a146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.276780] env[61907]: DEBUG nova.compute.manager [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1015.277017] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.277941] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64526de3-d367-47ce-9df7-4234438c86fe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.285564] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.285780] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c291956e-bec3-4136-b2c9-a597f7ef6ecd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.291715] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1015.291715] env[61907]: value = "task-1244295" [ 1015.291715] env[61907]: _type = "Task" [ 1015.291715] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.299303] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244295, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.423709] env[61907]: DEBUG nova.compute.manager [req-06dca27c-cd51-4344-ab6a-228434d821b9 req-2306c1ff-94ef-483a-b639-c51f022ca59a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Received event network-vif-deleted-bf950e7a-fdf3-453e-8bda-0ccca9cc8c98 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1015.423810] env[61907]: INFO nova.compute.manager [req-06dca27c-cd51-4344-ab6a-228434d821b9 req-2306c1ff-94ef-483a-b639-c51f022ca59a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Neutron deleted interface bf950e7a-fdf3-453e-8bda-0ccca9cc8c98; detaching it from the instance and deleting it from the info cache [ 1015.424035] env[61907]: DEBUG nova.network.neutron [req-06dca27c-cd51-4344-ab6a-228434d821b9 req-2306c1ff-94ef-483a-b639-c51f022ca59a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.557986] env[61907]: DEBUG nova.scheduler.client.report [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1015.618519] env[61907]: DEBUG nova.network.neutron [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.684515] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.684841] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance network_info: |[{"id": "cd2253b4-0244-45b6-80cb-b483ff82a146", "address": "fa:16:3e:35:ff:d4", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2253b4-02", "ovs_interfaceid": "cd2253b4-0244-45b6-80cb-b483ff82a146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1015.685288] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:35:ff:d4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '53955f0e-c162-4cef-8bd5-335b369c36b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cd2253b4-0244-45b6-80cb-b483ff82a146', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.692972] env[61907]: DEBUG oslo.service.loopingcall [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.695221] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.695460] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5b4cc17-d822-49f0-bffc-520ed9216f1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.714488] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.714488] env[61907]: value = "task-1244296" [ 1015.714488] env[61907]: _type = "Task" [ 1015.714488] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.721799] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244296, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.802073] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244295, 'name': PowerOffVM_Task, 'duration_secs': 0.198989} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.802450] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.802571] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.802918] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-244eeec9-aa13-42b5-972d-dc05d92a3155 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.858515] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.858744] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.859322] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Deleting the datastore file [datastore2] 6046a82d-f73d-4369-8208-c5a2e5c77431 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.859322] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99b10dbe-b013-447f-9aed-a9e07f607518 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.864972] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for the task: (returnval){ [ 1015.864972] env[61907]: value = "task-1244298" [ 1015.864972] env[61907]: _type = "Task" [ 1015.864972] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.872356] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244298, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.896625] env[61907]: DEBUG nova.network.neutron [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.926922] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-014825a6-478d-451d-af7c-ea0bb6765147 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.936104] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e529af-409f-461d-bee9-c3ccde10752c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.960865] env[61907]: DEBUG nova.compute.manager [req-06dca27c-cd51-4344-ab6a-228434d821b9 req-2306c1ff-94ef-483a-b639-c51f022ca59a service nova] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Detach interface failed, port_id=bf950e7a-fdf3-453e-8bda-0ccca9cc8c98, reason: Instance fcde07bf-858a-4377-b27c-4f17356306af could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1016.063333] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.684s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.081016] env[61907]: INFO nova.scheduler.client.report [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted allocations for instance 23526967-cd8f-4581-b9c5-1c270d385163 [ 1016.121039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.224046] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244296, 'name': CreateVM_Task, 'duration_secs': 0.292213} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.224285] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1016.224938] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.225131] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.226052] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1016.226052] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30c75c6a-3c18-4db8-9e15-cc3a5eaca599 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.230312] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1016.230312] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5216f17d-4e9d-ff2e-907b-0921b3fa66a3" [ 1016.230312] env[61907]: _type = "Task" [ 1016.230312] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.237679] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5216f17d-4e9d-ff2e-907b-0921b3fa66a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.375697] env[61907]: DEBUG oslo_vmware.api [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Task: {'id': task-1244298, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147588} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.375991] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.376310] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.376514] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.376700] env[61907]: INFO nova.compute.manager [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1016.376955] env[61907]: DEBUG oslo.service.loopingcall [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.377205] env[61907]: DEBUG nova.compute.manager [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1016.377316] env[61907]: DEBUG nova.network.neutron [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.399488] env[61907]: INFO nova.compute.manager [-] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Took 1.30 seconds to deallocate network for instance. [ 1016.592021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-87947c76-9f67-4045-8412-5603ef93915e tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "23526967-cd8f-4581-b9c5-1c270d385163" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.015s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.741124] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5216f17d-4e9d-ff2e-907b-0921b3fa66a3, 'name': SearchDatastore_Task, 'duration_secs': 0.00871} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.741459] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.741704] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.741940] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.742110] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.742297] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.742557] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-04e7d705-d5c2-46fd-91a8-60cefe37c80c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.751123] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.751318] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.752038] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7f6708d-04e6-43e2-9c69-4e82acca2b86 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.756834] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1016.756834] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52256805-7282-71ca-98e0-4a50f5d226b7" [ 1016.756834] env[61907]: _type = "Task" [ 1016.756834] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.763728] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52256805-7282-71ca-98e0-4a50f5d226b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.906465] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.906899] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.907287] env[61907]: DEBUG nova.objects.instance [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lazy-loading 'resources' on Instance uuid fcde07bf-858a-4377-b27c-4f17356306af {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.118144] env[61907]: DEBUG nova.network.neutron [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.267627] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52256805-7282-71ca-98e0-4a50f5d226b7, 'name': SearchDatastore_Task, 'duration_secs': 0.00746} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.268423] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b51e70e8-4996-4c8e-9812-70d744f62412 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.273525] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1017.273525] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5203c057-4221-ea29-b90f-7f917f72f698" [ 1017.273525] env[61907]: _type = "Task" [ 1017.273525] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.281387] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5203c057-4221-ea29-b90f-7f917f72f698, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.450032] env[61907]: DEBUG nova.compute.manager [req-8cc88d75-0189-4ae5-be28-9e203235da48 req-d48164ff-445a-48c2-b669-d01dd724df4c service nova] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Received event network-vif-deleted-5e64c307-2d27-4555-9376-fd3d08be3fc0 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1017.506829] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dcd1453-30da-458c-bf08-29fa31aa7a9f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.513998] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b6997b-1ce8-44ee-b9ae-52f76b4f213d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.544281] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b1d678-2567-43c8-9e00-6757b2e8140e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.551338] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197b692d-1304-4c30-8b0a-c0598d76d7ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.563991] env[61907]: DEBUG nova.compute.provider_tree [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.574370] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.574647] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.574778] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1017.574970] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.575160] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.576946] env[61907]: INFO nova.compute.manager [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Terminating instance [ 1017.620999] env[61907]: INFO nova.compute.manager [-] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Took 1.24 seconds to deallocate network for instance. [ 1017.637016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70c21c08-61de-4613-9fb1-c835d1d70528 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.655791] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1017.784779] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5203c057-4221-ea29-b90f-7f917f72f698, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.784779] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.784779] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] a84ffe50-2bd1-46f3-bbfd-e95f8dc23904/a84ffe50-2bd1-46f3-bbfd-e95f8dc23904.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.785049] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39fed0b6-1415-446a-83a6-9e47c88f0f6b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.791934] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1017.791934] env[61907]: value = "task-1244299" [ 1017.791934] env[61907]: _type = "Task" [ 1017.791934] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.799320] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244299, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.067019] env[61907]: DEBUG nova.scheduler.client.report [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1018.080869] env[61907]: DEBUG nova.compute.manager [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1018.081186] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1018.082736] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f8ab0c7-f09b-459d-9563-867d930c9227 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.091604] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.091895] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-445f746a-7a75-4252-bb40-77ee5e73da76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.098832] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 1018.098832] env[61907]: value = "task-1244300" [ 1018.098832] env[61907]: _type = "Task" [ 1018.098832] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.108434] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244300, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.127529] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.161232] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1018.161574] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9acbf37-160e-4141-ba24-ea30e598ea54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.169520] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1018.169520] env[61907]: value = "task-1244301" [ 1018.169520] env[61907]: _type = "Task" [ 1018.169520] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.180411] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244301, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.304026] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244299, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438225} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.304026] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] a84ffe50-2bd1-46f3-bbfd-e95f8dc23904/a84ffe50-2bd1-46f3-bbfd-e95f8dc23904.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1018.304026] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1018.304377] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a8d3add-7158-42f0-adf1-05c4dcc7c9ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.313029] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1018.313029] env[61907]: value = "task-1244302" [ 1018.313029] env[61907]: _type = "Task" [ 1018.313029] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.320859] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244302, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.572222] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.574795] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.447s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.575227] env[61907]: DEBUG nova.objects.instance [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lazy-loading 'resources' on Instance uuid 6046a82d-f73d-4369-8208-c5a2e5c77431 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.590494] env[61907]: INFO nova.scheduler.client.report [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Deleted allocations for instance fcde07bf-858a-4377-b27c-4f17356306af [ 1018.608938] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244300, 'name': PowerOffVM_Task, 'duration_secs': 0.204445} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.609260] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.609451] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1018.609765] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a497540-d5c3-4520-8085-e41cad2a7b76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.675558] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1018.675791] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1018.675975] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleting the datastore file [datastore2] 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1018.676822] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d70d34fc-26a4-4c96-8d5b-5ff19d1a9087 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.681401] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244301, 'name': PowerOffVM_Task, 'duration_secs': 0.215273} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.681941] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1018.682156] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1018.686276] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for the task: (returnval){ [ 1018.686276] env[61907]: value = "task-1244304" [ 1018.686276] env[61907]: _type = "Task" [ 1018.686276] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.694560] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244304, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.823815] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244302, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074779} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.825061] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.825311] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baea8be2-3025-4bcb-adc1-7fe7b1165cd1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.847480] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] a84ffe50-2bd1-46f3-bbfd-e95f8dc23904/a84ffe50-2bd1-46f3-bbfd-e95f8dc23904.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.847774] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3068c5f3-c08d-4f90-b551-12240f554281 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.866252] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1018.866252] env[61907]: value = "task-1244305" [ 1018.866252] env[61907]: _type = "Task" [ 1018.866252] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.873725] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.098137] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b8a918d7-600b-4691-9556-bfe87030dc65 tempest-AttachInterfacesTestJSON-210675951 tempest-AttachInterfacesTestJSON-210675951-project-member] Lock "fcde07bf-858a-4377-b27c-4f17356306af" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.613s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.160361] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605dd9f7-c9c2-4c18-8e98-1ce4b83faf2b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.167736] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84f5fa08-ea5a-42cd-a9ca-b69840d451d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.200399] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.200727] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.200901] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.201114] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.201279] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.201440] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.201658] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.201826] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.202008] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.202193] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.202382] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.207498] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35cb6f48-d62f-40ed-8118-e1e6d0cf6d76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.220519] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eda4d87-6f27-431a-91ae-52bfd551e581 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.229816] env[61907]: DEBUG oslo_vmware.api [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Task: {'id': task-1244304, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153166} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.231039] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1019.231250] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1019.231437] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1019.231620] env[61907]: INFO nova.compute.manager [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1019.231900] env[61907]: DEBUG oslo.service.loopingcall [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.233028] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3408586-e763-456e-a708-98f06f03fe0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.236611] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1019.236611] env[61907]: value = "task-1244306" [ 1019.236611] env[61907]: _type = "Task" [ 1019.236611] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.236824] env[61907]: DEBUG nova.compute.manager [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1019.236909] env[61907]: DEBUG nova.network.neutron [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1019.249516] env[61907]: DEBUG nova.compute.provider_tree [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.257597] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244306, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.386309] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244305, 'name': ReconfigVM_Task, 'duration_secs': 0.265755} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.386309] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfigured VM instance instance-0000005f to attach disk [datastore1] a84ffe50-2bd1-46f3-bbfd-e95f8dc23904/a84ffe50-2bd1-46f3-bbfd-e95f8dc23904.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.386309] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c56d5e5-a475-4749-8114-293e388fe638 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.395226] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1019.395226] env[61907]: value = "task-1244307" [ 1019.395226] env[61907]: _type = "Task" [ 1019.395226] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.410563] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244307, 'name': Rename_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.730520] env[61907]: DEBUG nova.compute.manager [req-b713d51c-6800-4323-8a6a-307b2705c79b req-f1625566-65f5-4d32-a21a-b70e27ce8685 service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Received event network-vif-deleted-9a182520-96f9-434d-844a-fe3e6cc88fbc {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1019.730759] env[61907]: INFO nova.compute.manager [req-b713d51c-6800-4323-8a6a-307b2705c79b req-f1625566-65f5-4d32-a21a-b70e27ce8685 service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Neutron deleted interface 9a182520-96f9-434d-844a-fe3e6cc88fbc; detaching it from the instance and deleting it from the info cache [ 1019.730982] env[61907]: DEBUG nova.network.neutron [req-b713d51c-6800-4323-8a6a-307b2705c79b req-f1625566-65f5-4d32-a21a-b70e27ce8685 service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.748782] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244306, 'name': ReconfigVM_Task, 'duration_secs': 0.296068} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.749105] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1019.753569] env[61907]: DEBUG nova.scheduler.client.report [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1019.906548] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244307, 'name': Rename_Task, 'duration_secs': 0.203626} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.906896] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1019.907167] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e27f690-5808-413d-a295-bcc6d2202aee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.913732] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1019.913732] env[61907]: value = "task-1244308" [ 1019.913732] env[61907]: _type = "Task" [ 1019.913732] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.926111] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244308, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.206410] env[61907]: DEBUG nova.network.neutron [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.233557] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c426a233-2899-4eee-9e4e-396978b4008b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.242679] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b7ddd7-6702-4725-adaa-7286db41b9b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.259950] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.260220] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.260393] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.260596] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.260774] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.260965] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.261344] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.261585] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.261811] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.262041] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.262276] env[61907]: DEBUG nova.virt.hardware [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.268199] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfiguring VM instance instance-0000003f to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1020.274049] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.276077] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dee8f0c-5352-4909-95f4-ed99c3bae712 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.289558] env[61907]: DEBUG nova.compute.manager [req-b713d51c-6800-4323-8a6a-307b2705c79b req-f1625566-65f5-4d32-a21a-b70e27ce8685 service nova] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Detach interface failed, port_id=9a182520-96f9-434d-844a-fe3e6cc88fbc, reason: Instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1020.295627] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1020.295627] env[61907]: value = "task-1244309" [ 1020.295627] env[61907]: _type = "Task" [ 1020.295627] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.303262] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244309, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.312730] env[61907]: INFO nova.scheduler.client.report [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Deleted allocations for instance 6046a82d-f73d-4369-8208-c5a2e5c77431 [ 1020.425089] env[61907]: DEBUG oslo_vmware.api [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244308, 'name': PowerOnVM_Task, 'duration_secs': 0.47529} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.425386] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1020.425578] env[61907]: INFO nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Took 7.93 seconds to spawn the instance on the hypervisor. [ 1020.425768] env[61907]: DEBUG nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1020.426612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bcc6df-d1b7-4b29-a446-cc8dbf6eefc8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.710255] env[61907]: INFO nova.compute.manager [-] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Took 1.47 seconds to deallocate network for instance. [ 1020.808455] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244309, 'name': ReconfigVM_Task, 'duration_secs': 0.1707} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.808593] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfigured VM instance instance-0000003f to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1020.811251] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2648f62a-3142-4cf3-b798-8144eae820db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.836988] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1020.837560] env[61907]: DEBUG oslo_concurrency.lockutils [None req-494acee9-16a0-4755-813d-c2bb358dd725 tempest-InstanceActionsNegativeTestJSON-1603910721 tempest-InstanceActionsNegativeTestJSON-1603910721-project-member] Lock "6046a82d-f73d-4369-8208-c5a2e5c77431" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.069s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.838478] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60127456-4f70-4ffe-8a27-11332d5656f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.859036] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1020.859036] env[61907]: value = "task-1244310" [ 1020.859036] env[61907]: _type = "Task" [ 1020.859036] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.869279] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244310, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.942544] env[61907]: INFO nova.compute.manager [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Took 16.23 seconds to build instance. [ 1021.216561] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.216870] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.217064] env[61907]: DEBUG nova.objects.instance [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lazy-loading 'resources' on Instance uuid 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.370062] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244310, 'name': ReconfigVM_Task, 'duration_secs': 0.263551} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.370062] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Reconfigured VM instance instance-0000003f to attach disk [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0/92f27804-8974-40c4-9663-b2b72f0bb8e0.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1021.370062] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.445699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7e03ea66-8c00-40d4-8fd5-47563559f923 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.746s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.776492] env[61907]: DEBUG nova.compute.manager [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Received event network-changed-cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1021.776713] env[61907]: DEBUG nova.compute.manager [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Refreshing instance network info cache due to event network-changed-cd2253b4-0244-45b6-80cb-b483ff82a146. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1021.776930] env[61907]: DEBUG oslo_concurrency.lockutils [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] Acquiring lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.777093] env[61907]: DEBUG oslo_concurrency.lockutils [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] Acquired lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.777265] env[61907]: DEBUG nova.network.neutron [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Refreshing network info cache for port cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1021.816758] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7921844-5b16-4fc2-8a28-34d73911991b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.824295] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b81b075-9a9a-4985-833c-5926467ac0d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.857363] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63aaa192-06cf-4053-8469-237108bab212 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.864885] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afdd587e-7902-4454-8341-d6ee421c0cf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.881041] env[61907]: DEBUG nova.compute.provider_tree [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1021.884089] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f29e75-e889-4f94-a631-684ea2c6b2d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.907074] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-205e67a8-d914-4176-89a2-bfc600d3b609 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.926582] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.388237] env[61907]: DEBUG nova.scheduler.client.report [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1022.491898] env[61907]: DEBUG nova.network.neutron [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Port f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1022.540210] env[61907]: DEBUG nova.network.neutron [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updated VIF entry in instance network info cache for port cd2253b4-0244-45b6-80cb-b483ff82a146. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1022.540590] env[61907]: DEBUG nova.network.neutron [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating instance_info_cache with network_info: [{"id": "cd2253b4-0244-45b6-80cb-b483ff82a146", "address": "fa:16:3e:35:ff:d4", "network": {"id": "269802d8-0fcd-48e5-93d1-5cc3e1d6a50c", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1576379830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.131", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a85baf8058442b39a07436a046e1ebe", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "53955f0e-c162-4cef-8bd5-335b369c36b6", "external-id": "nsx-vlan-transportzone-623", "segmentation_id": 623, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcd2253b4-02", "ovs_interfaceid": "cd2253b4-0244-45b6-80cb-b483ff82a146", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.896111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.679s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.921485] env[61907]: INFO nova.scheduler.client.report [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Deleted allocations for instance 2ab93ab9-bf69-4525-8df6-eef83dd24bc1 [ 1023.043584] env[61907]: DEBUG oslo_concurrency.lockutils [req-ec1ea995-a7fb-4b6e-9aac-1efdf97fbf67 req-2425c6bb-1752-431a-8683-c5bcdea63232 service nova] Releasing lock "refresh_cache-a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.429964] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9749ee73-fb2e-41ad-8409-4ab0cee744e3 tempest-ServerRescueNegativeTestJSON-2062051343 tempest-ServerRescueNegativeTestJSON-2062051343-project-member] Lock "2ab93ab9-bf69-4525-8df6-eef83dd24bc1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.855s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.520021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.520021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.520021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.582549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.582549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.582549] env[61907]: DEBUG nova.network.neutron [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.482572] env[61907]: DEBUG nova.network.neutron [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.868718] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.868977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.913897] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "03053d82-188d-45f1-9123-240c50d25ea2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.914158] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.952148] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.952148] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.985320] env[61907]: DEBUG oslo_concurrency.lockutils [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.373932] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1026.417022] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1026.454276] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1026.522055] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69aef62c-df86-4654-9f93-510cd5fc41ac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.548154] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7884f4e-33e2-49d9-b2dc-14b09658c08c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.556601] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1026.906316] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.906623] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.908603] env[61907]: INFO nova.compute.claims [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.941158] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.982456] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.064308] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1027.064686] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dfe2e92-b5d1-40a7-9246-feb421b16afa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.072867] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1027.072867] env[61907]: value = "task-1244311" [ 1027.072867] env[61907]: _type = "Task" [ 1027.072867] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.081861] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244311, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.583571] env[61907]: DEBUG oslo_vmware.api [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244311, 'name': PowerOnVM_Task, 'duration_secs': 0.417009} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.583858] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1027.584900] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-1846c7df-0572-4db1-b30d-3ca94af4ce06 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance '92f27804-8974-40c4-9663-b2b72f0bb8e0' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1027.717928] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.718299] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.043142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af361585-6d9c-4a7b-84ea-4bf1650579f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.052696] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042683c8-b56e-40d8-b790-0e81565066f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.089012] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b8b911-fa63-42ad-83d6-61bf00aacad9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.098609] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae5e494d-71cb-4c2f-bc1c-6461173d7f9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.112871] env[61907]: DEBUG nova.compute.provider_tree [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1028.220578] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1028.615671] env[61907]: DEBUG nova.scheduler.client.report [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1028.747199] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.126037] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.218s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.126037] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1029.130444] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.189s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.136031] env[61907]: INFO nova.compute.claims [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.641528] env[61907]: DEBUG nova.compute.utils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.645029] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1029.645243] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.691571] env[61907]: DEBUG nova.policy [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5630bfcb4a004fcfab22c9812ddf3b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76a13ea608d401780c4ef909366c9f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.840911] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.841197] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.841440] env[61907]: DEBUG nova.compute.manager [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Going to confirm migration 4 {{(pid=61907) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1029.968732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "b2c5689e-f930-4548-b061-fd2974d44575" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.968974] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.993183] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Successfully created port: 60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1030.148607] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1030.309323] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d6bf9f8-8eba-4367-a14c-e39ab118d126 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.317679] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e6da9a-faa6-4e55-b78b-92a566589b6b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.351043] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae56d85-daa0-4e7e-aa51-604e38f661f9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.360269] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08095555-80c8-4290-ab0e-ce61317ad3c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.374922] env[61907]: DEBUG nova.compute.provider_tree [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.396997] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.397167] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.397384] env[61907]: DEBUG nova.network.neutron [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1030.400747] env[61907]: DEBUG nova.objects.instance [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'info_cache' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1030.471258] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1030.878404] env[61907]: DEBUG nova.scheduler.client.report [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1030.998058] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.161646] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1031.185630] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1031.185862] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1031.186035] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1031.186229] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1031.186395] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1031.186660] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1031.186883] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1031.187062] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1031.187237] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1031.187429] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1031.187630] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1031.188493] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0318b7-e100-4b00-b70b-fbf10b8cc639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.196313] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c96b11da-3f73-427d-89f1-1302ff9e00c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.387280] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.387280] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1031.391641] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.409s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.393143] env[61907]: INFO nova.compute.claims [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.408045] env[61907]: DEBUG nova.compute.manager [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Received event network-vif-plugged-60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1031.408119] env[61907]: DEBUG oslo_concurrency.lockutils [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] Acquiring lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.408310] env[61907]: DEBUG oslo_concurrency.lockutils [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.408478] env[61907]: DEBUG oslo_concurrency.lockutils [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.408648] env[61907]: DEBUG nova.compute.manager [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] No waiting events found dispatching network-vif-plugged-60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1031.408815] env[61907]: WARNING nova.compute.manager [req-b6248341-ce91-48a1-a2ea-80df857a3d7b req-3a5fb31a-9cc5-4b4f-92a1-67518daf7464 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Received unexpected event network-vif-plugged-60a000f9-0086-46cb-bf77-ed7e0b552d3b for instance with vm_state building and task_state spawning. [ 1031.584679] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Successfully updated port: 60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.717853] env[61907]: DEBUG nova.network.neutron [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [{"id": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "address": "fa:16:3e:93:5f:a5", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf5ac28f2-22", "ovs_interfaceid": "f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.898837] env[61907]: DEBUG nova.compute.utils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.904843] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1031.904843] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1031.944302] env[61907]: DEBUG nova.policy [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5630bfcb4a004fcfab22c9812ddf3b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76a13ea608d401780c4ef909366c9f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1032.090024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.090024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.090024] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.182335] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Successfully created port: c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.220240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-92f27804-8974-40c4-9663-b2b72f0bb8e0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.220480] env[61907]: DEBUG nova.objects.instance [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'migration_context' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.405339] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1032.508522] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.513901] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.522390] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e736ddac-d824-43b4-8d71-591adf04edf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.532739] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b86bea63-5556-40a3-a018-4b6e261a1192 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.563048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab2bca8-4262-4652-adea-851b8a97df6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.571512] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3341b784-ef58-4bae-b83a-16364281f61c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.584762] env[61907]: DEBUG nova.compute.provider_tree [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.620206] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.723365] env[61907]: DEBUG nova.objects.base [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Object Instance<92f27804-8974-40c4-9663-b2b72f0bb8e0> lazy-loaded attributes: info_cache,migration_context {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1032.725124] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ecdeff2-8815-4ed5-b7d4-abf7a16e98cb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.747053] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad48d5af-8dec-405c-a896-cd0b07f46be5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.754837] env[61907]: DEBUG oslo_vmware.api [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1032.754837] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4af41-2c84-c74c-70ff-cb741b34c978" [ 1032.754837] env[61907]: _type = "Task" [ 1032.754837] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.762788] env[61907]: DEBUG oslo_vmware.api [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4af41-2c84-c74c-70ff-cb741b34c978, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.983102] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Updating instance_info_cache with network_info: [{"id": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "address": "fa:16:3e:a8:75:1c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a000f9-00", "ovs_interfaceid": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.087219] env[61907]: DEBUG nova.scheduler.client.report [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1033.265366] env[61907]: DEBUG oslo_vmware.api [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e4af41-2c84-c74c-70ff-cb741b34c978, 'name': SearchDatastore_Task, 'duration_secs': 0.009674} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.265658] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.416849] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1033.441071] env[61907]: DEBUG nova.compute.manager [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Received event network-changed-60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1033.441282] env[61907]: DEBUG nova.compute.manager [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Refreshing instance network info cache due to event network-changed-60a000f9-0086-46cb-bf77-ed7e0b552d3b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1033.441485] env[61907]: DEBUG oslo_concurrency.lockutils [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] Acquiring lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.443540] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.443767] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.443930] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.444131] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.444281] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.444444] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.444757] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.445015] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.445207] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.445378] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.445557] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.446811] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eebd5f1-4af2-4762-99dd-51fe883d498f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.455395] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94f12ea-ab42-4253-a343-80d326a5b083 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.485390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.485747] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Instance network_info: |[{"id": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "address": "fa:16:3e:a8:75:1c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a000f9-00", "ovs_interfaceid": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1033.486073] env[61907]: DEBUG oslo_concurrency.lockutils [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] Acquired lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.486267] env[61907]: DEBUG nova.network.neutron [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Refreshing network info cache for port 60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.487491] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a8:75:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b5a629f-6902-4d30-9278-74b443a8371d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '60a000f9-0086-46cb-bf77-ed7e0b552d3b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.494741] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Creating folder: Project (b76a13ea608d401780c4ef909366c9f8). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1033.495206] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d88ec8b1-2781-4987-bcea-e577ac736d66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.506811] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Created folder: Project (b76a13ea608d401780c4ef909366c9f8) in parent group-v268168. [ 1033.506995] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Creating folder: Instances. Parent ref: group-v268313. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1033.507238] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e939f017-5ec4-43f7-91e5-5d0d9abf979a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.515589] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Created folder: Instances in parent group-v268313. [ 1033.515812] env[61907]: DEBUG oslo.service.loopingcall [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.515992] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.516200] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fdf4c57d-b142-4fa4-9536-c47eb1a3da4e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.537019] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.537019] env[61907]: value = "task-1244314" [ 1033.537019] env[61907]: _type = "Task" [ 1033.537019] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.541507] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244314, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.592383] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.593125] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1033.596229] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.849s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.597793] env[61907]: INFO nova.compute.claims [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.681054] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Successfully updated port: c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1034.043814] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244314, 'name': CreateVM_Task, 'duration_secs': 0.330859} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.043964] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.044689] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.045819] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.045819] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.045819] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49f0eafc-47eb-4c7d-8acc-037f6672d61b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.049874] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1034.049874] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524f4990-ed12-9d80-4461-b00c2de5d82a" [ 1034.049874] env[61907]: _type = "Task" [ 1034.049874] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.057039] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524f4990-ed12-9d80-4461-b00c2de5d82a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.103044] env[61907]: DEBUG nova.compute.utils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1034.105683] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1034.105851] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1034.169635] env[61907]: DEBUG nova.policy [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5630bfcb4a004fcfab22c9812ddf3b35', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b76a13ea608d401780c4ef909366c9f8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1034.183017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.183182] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.183330] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1034.242665] env[61907]: DEBUG nova.network.neutron [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Updated VIF entry in instance network info cache for port 60a000f9-0086-46cb-bf77-ed7e0b552d3b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1034.243158] env[61907]: DEBUG nova.network.neutron [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Updating instance_info_cache with network_info: [{"id": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "address": "fa:16:3e:a8:75:1c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap60a000f9-00", "ovs_interfaceid": "60a000f9-0086-46cb-bf77-ed7e0b552d3b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.444339] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Successfully created port: 0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1034.560726] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524f4990-ed12-9d80-4461-b00c2de5d82a, 'name': SearchDatastore_Task, 'duration_secs': 0.008774} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.561060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.561303] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.561540] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.561694] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.561877] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.562169] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dceff498-4547-4b2b-b1ee-fef19854c7ea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.570262] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.570487] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.571203] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38ac0dd2-1138-42ef-a9b6-21facaa704d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.576043] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1034.576043] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d09b0f-f6ea-a30c-2ce0-d9ada6793ab8" [ 1034.576043] env[61907]: _type = "Task" [ 1034.576043] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.583634] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d09b0f-f6ea-a30c-2ce0-d9ada6793ab8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.607165] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1034.714730] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1034.746498] env[61907]: DEBUG oslo_concurrency.lockutils [req-9983ed1e-1b24-47c6-900f-46799b162314 req-a9d4e0d3-6539-4b28-ab97-8e1322a285e6 service nova] Releasing lock "refresh_cache-3aeeac7d-6642-450e-9c1a-e43d514389a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.757824] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9b4299-d987-45e1-8205-46aa6bb24663 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.764997] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54862417-3f32-4159-9aba-930df2059436 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.797587] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b680fcad-aeb9-45de-a628-a1c74d855f37 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.805356] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b292046-99c9-42cc-b6c9-94f45aba30cb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.818228] env[61907]: DEBUG nova.compute.provider_tree [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.883318] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Updating instance_info_cache with network_info: [{"id": "c5964c65-72ff-4a67-a68a-e746390660c4", "address": "fa:16:3e:f2:68:d0", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5964c65-72", "ovs_interfaceid": "c5964c65-72ff-4a67-a68a-e746390660c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.086777] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d09b0f-f6ea-a30c-2ce0-d9ada6793ab8, 'name': SearchDatastore_Task, 'duration_secs': 0.008443} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.087544] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d1f6925-e39e-47ec-92f1-4ddd9ca2d767 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.092532] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1035.092532] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524a263a-0012-eb40-b0b1-ff222641bcdb" [ 1035.092532] env[61907]: _type = "Task" [ 1035.092532] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.099929] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524a263a-0012-eb40-b0b1-ff222641bcdb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.321537] env[61907]: DEBUG nova.scheduler.client.report [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1035.385728] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.386412] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Instance network_info: |[{"id": "c5964c65-72ff-4a67-a68a-e746390660c4", "address": "fa:16:3e:f2:68:d0", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5964c65-72", "ovs_interfaceid": "c5964c65-72ff-4a67-a68a-e746390660c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1035.386593] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:68:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b5a629f-6902-4d30-9278-74b443a8371d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5964c65-72ff-4a67-a68a-e746390660c4', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.394058] env[61907]: DEBUG oslo.service.loopingcall [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.394284] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.394554] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6b3ebab-6693-4636-b5a6-89c40fde3bbe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.415496] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.415496] env[61907]: value = "task-1244315" [ 1035.415496] env[61907]: _type = "Task" [ 1035.415496] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.425330] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244315, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.472978] env[61907]: DEBUG nova.compute.manager [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Received event network-vif-plugged-c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1035.472978] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Acquiring lock "03053d82-188d-45f1-9123-240c50d25ea2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.474139] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Lock "03053d82-188d-45f1-9123-240c50d25ea2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.474139] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Lock "03053d82-188d-45f1-9123-240c50d25ea2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.474139] env[61907]: DEBUG nova.compute.manager [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] No waiting events found dispatching network-vif-plugged-c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1035.474139] env[61907]: WARNING nova.compute.manager [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Received unexpected event network-vif-plugged-c5964c65-72ff-4a67-a68a-e746390660c4 for instance with vm_state building and task_state spawning. [ 1035.474139] env[61907]: DEBUG nova.compute.manager [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Received event network-changed-c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1035.474139] env[61907]: DEBUG nova.compute.manager [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Refreshing instance network info cache due to event network-changed-c5964c65-72ff-4a67-a68a-e746390660c4. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1035.474384] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Acquiring lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.474468] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Acquired lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.474647] env[61907]: DEBUG nova.network.neutron [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Refreshing network info cache for port c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1035.513727] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1035.513940] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1035.604292] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524a263a-0012-eb40-b0b1-ff222641bcdb, 'name': SearchDatastore_Task, 'duration_secs': 0.008372} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.604614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.604935] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 3aeeac7d-6642-450e-9c1a-e43d514389a3/3aeeac7d-6642-450e-9c1a-e43d514389a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.605167] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-174081c8-fbf7-4400-9681-037c1a98be1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.612210] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1035.612210] env[61907]: value = "task-1244316" [ 1035.612210] env[61907]: _type = "Task" [ 1035.612210] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.620397] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1035.622420] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244316, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.645460] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.645585] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.645735] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.645929] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.646089] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.646810] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.646810] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.646810] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.646810] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.646997] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.647170] env[61907]: DEBUG nova.virt.hardware [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.648036] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17ff1f25-8c0c-4ad1-af16-7475c650fed3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.655192] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d69a46-a351-431c-a4ea-4255356caa1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.826163] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.826774] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1035.829673] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.832s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.832138] env[61907]: INFO nova.compute.claims [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1035.926576] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244315, 'name': CreateVM_Task, 'duration_secs': 0.29001} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.926845] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1035.927940] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.928141] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.928493] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.928864] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9fc3a90-ffc2-4aa8-918f-8292a63a3e7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.934978] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1035.934978] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52695ad8-2cf9-2e5b-8f79-8d071267b29a" [ 1035.934978] env[61907]: _type = "Task" [ 1035.934978] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.944525] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52695ad8-2cf9-2e5b-8f79-8d071267b29a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.953708] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Successfully updated port: 0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1036.122036] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244316, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423235} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.122544] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 3aeeac7d-6642-450e-9c1a-e43d514389a3/3aeeac7d-6642-450e-9c1a-e43d514389a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1036.122544] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.123313] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bd0c6fd-138b-41a6-be66-d35ca592aa4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.131742] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1036.131742] env[61907]: value = "task-1244317" [ 1036.131742] env[61907]: _type = "Task" [ 1036.131742] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.139666] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244317, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.200728] env[61907]: DEBUG nova.network.neutron [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Updated VIF entry in instance network info cache for port c5964c65-72ff-4a67-a68a-e746390660c4. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1036.201124] env[61907]: DEBUG nova.network.neutron [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Updating instance_info_cache with network_info: [{"id": "c5964c65-72ff-4a67-a68a-e746390660c4", "address": "fa:16:3e:f2:68:d0", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5964c65-72", "ovs_interfaceid": "c5964c65-72ff-4a67-a68a-e746390660c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.335235] env[61907]: DEBUG nova.compute.utils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.340060] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1036.340255] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1036.377901] env[61907]: DEBUG nova.policy [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8186074291264496984da491880a55b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f851fef407334ba2919877cfd4865435', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.447541] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52695ad8-2cf9-2e5b-8f79-8d071267b29a, 'name': SearchDatastore_Task, 'duration_secs': 0.045641} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.447541] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.447541] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.447541] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.447541] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.447541] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.447932] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b041bde-7fae-46a2-a6ce-d95582c817db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.456446] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.456584] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.456819] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1036.458079] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.458307] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1036.459078] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00b7f554-857c-4a5c-b366-35cec65bb411 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.467056] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1036.467056] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52381e21-af37-3791-8a66-dc84f1fc1577" [ 1036.467056] env[61907]: _type = "Task" [ 1036.467056] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.475222] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52381e21-af37-3791-8a66-dc84f1fc1577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.521046] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Didn't find any instances for network info cache update. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 1036.521278] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.521451] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.521603] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.521754] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1036.641684] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244317, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062927} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.641973] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.642759] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2126931e-d70f-41bb-bf1d-a93888e1bf1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.664808] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 3aeeac7d-6642-450e-9c1a-e43d514389a3/3aeeac7d-6642-450e-9c1a-e43d514389a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.665620] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Successfully created port: 6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.667679] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e38cabf-6c17-48d1-a6ae-07e83fd4767d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.688302] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1036.688302] env[61907]: value = "task-1244318" [ 1036.688302] env[61907]: _type = "Task" [ 1036.688302] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.698576] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244318, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.704143] env[61907]: DEBUG oslo_concurrency.lockutils [req-d699e62e-f50e-45de-bf6d-ff67372eb68b req-39a23442-fef3-4f66-a43d-cb08bab4ab40 service nova] Releasing lock "refresh_cache-03053d82-188d-45f1-9123-240c50d25ea2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.841615] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1036.976168] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52381e21-af37-3791-8a66-dc84f1fc1577, 'name': SearchDatastore_Task, 'duration_secs': 0.008522} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.979027] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d16a4c7-e9fa-4a2f-9ea6-d411d4805cbc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.981489] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae30e6d-94d8-401f-bfb2-5089d3f0828f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.985135] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1036.985135] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fa002f-a834-471a-a864-2236a4787232" [ 1036.985135] env[61907]: _type = "Task" [ 1036.985135] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.990834] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde28e02-fba3-4e3b-9219-8e3b10e99068 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.994580] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1036.999345] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52fa002f-a834-471a-a864-2236a4787232, 'name': SearchDatastore_Task, 'duration_secs': 0.008238} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.000152] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.000398] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 03053d82-188d-45f1-9123-240c50d25ea2/03053d82-188d-45f1-9123-240c50d25ea2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1037.000625] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33a83870-7909-4e92-8034-523833045fa1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.028184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.032031] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d9a835-9c01-4f49-8dd3-bfb5226ba9c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.036266] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1037.036266] env[61907]: value = "task-1244319" [ 1037.036266] env[61907]: _type = "Task" [ 1037.036266] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.042098] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29f6a14c-1ec7-4c7a-87d4-315da41de91d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.048445] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244319, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.057780] env[61907]: DEBUG nova.compute.provider_tree [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.152026] env[61907]: DEBUG nova.network.neutron [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Updating instance_info_cache with network_info: [{"id": "0e9518d1-76b7-4257-829c-e21a49a48015", "address": "fa:16:3e:7e:08:9c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e9518d1-76", "ovs_interfaceid": "0e9518d1-76b7-4257-829c-e21a49a48015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.198744] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244318, 'name': ReconfigVM_Task, 'duration_secs': 0.278308} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.199133] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 3aeeac7d-6642-450e-9c1a-e43d514389a3/3aeeac7d-6642-450e-9c1a-e43d514389a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.199703] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-12b72c14-8d2b-4a4a-8f76-10b42b05c5ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.205720] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1037.205720] env[61907]: value = "task-1244320" [ 1037.205720] env[61907]: _type = "Task" [ 1037.205720] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.215782] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244320, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.504604] env[61907]: DEBUG nova.compute.manager [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Received event network-vif-plugged-0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1037.505043] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Acquiring lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1037.505043] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.505103] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.505265] env[61907]: DEBUG nova.compute.manager [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] No waiting events found dispatching network-vif-plugged-0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1037.505437] env[61907]: WARNING nova.compute.manager [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Received unexpected event network-vif-plugged-0e9518d1-76b7-4257-829c-e21a49a48015 for instance with vm_state building and task_state spawning. [ 1037.505607] env[61907]: DEBUG nova.compute.manager [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Received event network-changed-0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1037.505812] env[61907]: DEBUG nova.compute.manager [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Refreshing instance network info cache due to event network-changed-0e9518d1-76b7-4257-829c-e21a49a48015. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1037.506311] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Acquiring lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.546570] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244319, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487982} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.546751] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 03053d82-188d-45f1-9123-240c50d25ea2/03053d82-188d-45f1-9123-240c50d25ea2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.546973] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.547247] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8d2c3158-dec4-4854-98ff-07caadd5ba4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.554855] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1037.554855] env[61907]: value = "task-1244321" [ 1037.554855] env[61907]: _type = "Task" [ 1037.554855] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.561990] env[61907]: DEBUG nova.scheduler.client.report [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1037.565022] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244321, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.657406] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.657739] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Instance network_info: |[{"id": "0e9518d1-76b7-4257-829c-e21a49a48015", "address": "fa:16:3e:7e:08:9c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e9518d1-76", "ovs_interfaceid": "0e9518d1-76b7-4257-829c-e21a49a48015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1037.658072] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Acquired lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.658281] env[61907]: DEBUG nova.network.neutron [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Refreshing network info cache for port 0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1037.659940] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7e:08:9c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b5a629f-6902-4d30-9278-74b443a8371d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e9518d1-76b7-4257-829c-e21a49a48015', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.667236] env[61907]: DEBUG oslo.service.loopingcall [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.669983] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1037.670465] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d4c8eb0-55cd-48d7-a5ed-e54fdb2e0069 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.690415] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.690415] env[61907]: value = "task-1244322" [ 1037.690415] env[61907]: _type = "Task" [ 1037.690415] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.699802] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244322, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.713679] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244320, 'name': Rename_Task, 'duration_secs': 0.151226} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.715916] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1037.716456] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-40c559b7-21ef-4272-a5eb-4b5ab3f98823 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.723121] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1037.723121] env[61907]: value = "task-1244323" [ 1037.723121] env[61907]: _type = "Task" [ 1037.723121] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.730999] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244323, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.859016] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1037.888040] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1037.888202] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1037.888305] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1037.888505] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1037.888663] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1037.888820] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1037.889047] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1037.889223] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1037.889403] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1037.889575] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1037.889757] env[61907]: DEBUG nova.virt.hardware [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1037.890649] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b480cb6-c918-4b11-b061-ae0767f379d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.899275] env[61907]: DEBUG nova.network.neutron [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Updated VIF entry in instance network info cache for port 0e9518d1-76b7-4257-829c-e21a49a48015. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1037.899657] env[61907]: DEBUG nova.network.neutron [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Updating instance_info_cache with network_info: [{"id": "0e9518d1-76b7-4257-829c-e21a49a48015", "address": "fa:16:3e:7e:08:9c", "network": {"id": "7b17295b-6e52-4f34-a3b5-ed8294b089c7", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-781336175-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b76a13ea608d401780c4ef909366c9f8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b5a629f-6902-4d30-9278-74b443a8371d", "external-id": "nsx-vlan-transportzone-185", "segmentation_id": 185, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e9518d1-76", "ovs_interfaceid": "0e9518d1-76b7-4257-829c-e21a49a48015", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.902090] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac7c714-5646-4094-920d-ffce824c2407 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.064111] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244321, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061756} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.064416] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.065242] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e81be3-6cc6-467d-b6e1-c7c6cecf4199 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.072390] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.072931] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1038.075512] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.810s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.102971] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 03053d82-188d-45f1-9123-240c50d25ea2/03053d82-188d-45f1-9123-240c50d25ea2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.104042] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-06bf03d3-75bb-4c95-bf70-b82607cfd532 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.128714] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1038.128714] env[61907]: value = "task-1244324" [ 1038.128714] env[61907]: _type = "Task" [ 1038.128714] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.137299] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.201546] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244322, 'name': CreateVM_Task, 'duration_secs': 0.405534} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.201744] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1038.202628] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.202920] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.203276] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1038.203589] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3622eb7-ece0-4235-84b8-8ccd2d784a51 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.209434] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1038.209434] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e04b05-af18-8c4c-8996-4b924006e537" [ 1038.209434] env[61907]: _type = "Task" [ 1038.209434] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.218515] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e04b05-af18-8c4c-8996-4b924006e537, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.232701] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244323, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.274279] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Successfully updated port: 6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.407655] env[61907]: DEBUG oslo_concurrency.lockutils [req-931e89ce-3083-4841-a742-f1295ea43cf6 req-bb7ca764-efe7-4d84-9f71-37b3aefdbec8 service nova] Releasing lock "refresh_cache-aaa04afe-15d8-4912-a51e-3d72bc70da9f" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.579289] env[61907]: DEBUG nova.compute.utils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.583773] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1038.583773] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.626909] env[61907]: DEBUG nova.policy [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.637213] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244324, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.695110] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ade5ca-d35b-4ef3-b6a9-a4cf944d360b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.701806] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3843440e-9f19-4778-b7ea-7086d46c88a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.736726] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953f45c6-41b2-44cc-bf69-40095c2fad33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.744035] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244323, 'name': PowerOnVM_Task, 'duration_secs': 0.510948} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.748999] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1038.749255] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Took 7.59 seconds to spawn the instance on the hypervisor. [ 1038.749446] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1038.749745] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e04b05-af18-8c4c-8996-4b924006e537, 'name': SearchDatastore_Task, 'duration_secs': 0.009522} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.750506] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7296cc99-cd28-4176-a3fa-f7b2cd98d682 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.752871] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.753117] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.753369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.753523] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.753708] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.754907] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5383e107-b715-45d9-af6b-e096762c0714 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.758829] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64e3f99d-8c0f-444d-9b0b-239069a739fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.774620] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.774818] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1038.775755] env[61907]: DEBUG nova.compute.provider_tree [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.777034] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e0d84dc-560c-4075-98e4-0ef511f91bf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.779469] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.779570] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.779710] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1038.784283] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1038.784283] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529342c4-af6c-3156-ba2d-81c05bea89fa" [ 1038.784283] env[61907]: _type = "Task" [ 1038.784283] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.792142] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529342c4-af6c-3156-ba2d-81c05bea89fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.882872] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Successfully created port: b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.084541] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1039.138561] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244324, 'name': ReconfigVM_Task, 'duration_secs': 0.790878} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.139660] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 03053d82-188d-45f1-9123-240c50d25ea2/03053d82-188d-45f1-9123-240c50d25ea2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.139660] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fce118a3-561a-4998-8fcb-e7036a736f98 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.146997] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1039.146997] env[61907]: value = "task-1244325" [ 1039.146997] env[61907]: _type = "Task" [ 1039.146997] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.154390] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244325, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.282026] env[61907]: DEBUG nova.scheduler.client.report [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1039.291719] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Took 12.41 seconds to build instance. [ 1039.298189] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529342c4-af6c-3156-ba2d-81c05bea89fa, 'name': SearchDatastore_Task, 'duration_secs': 0.008831} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.299466] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c5016a7-77f9-46fb-8f9b-84d385b370dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.305160] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1039.305160] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52903409-731e-8586-1a4a-ff8d222ec026" [ 1039.305160] env[61907]: _type = "Task" [ 1039.305160] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.313854] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52903409-731e-8586-1a4a-ff8d222ec026, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.317561] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1039.454994] env[61907]: DEBUG nova.network.neutron [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.533984] env[61907]: DEBUG nova.compute.manager [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Received event network-vif-plugged-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1039.534156] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.534333] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.534491] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.534672] env[61907]: DEBUG nova.compute.manager [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] No waiting events found dispatching network-vif-plugged-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1039.534840] env[61907]: WARNING nova.compute.manager [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Received unexpected event network-vif-plugged-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 for instance with vm_state building and task_state spawning. [ 1039.535015] env[61907]: DEBUG nova.compute.manager [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Received event network-changed-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1039.535354] env[61907]: DEBUG nova.compute.manager [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Refreshing instance network info cache due to event network-changed-6e35dd41-7b4c-44e2-8076-321b9b9bfc61. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1039.535535] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.656040] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244325, 'name': Rename_Task, 'duration_secs': 0.144683} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.656303] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.656573] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7184b2d8-4c05-499d-8924-89451cfaa9d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.662720] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1039.662720] env[61907]: value = "task-1244326" [ 1039.662720] env[61907]: _type = "Task" [ 1039.662720] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.672192] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244326, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.794011] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.925s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.815683] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52903409-731e-8586-1a4a-ff8d222ec026, 'name': SearchDatastore_Task, 'duration_secs': 0.008744} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.815956] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.816229] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] aaa04afe-15d8-4912-a51e-3d72bc70da9f/aaa04afe-15d8-4912-a51e-3d72bc70da9f.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1039.816502] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e91e731-cd05-4797-81ed-61dfabb7b302 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.823057] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1039.823057] env[61907]: value = "task-1244327" [ 1039.823057] env[61907]: _type = "Task" [ 1039.823057] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.830842] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244327, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.957126] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.957555] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Instance network_info: |[{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1039.957933] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.958155] env[61907]: DEBUG nova.network.neutron [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Refreshing network info cache for port 6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.959552] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:03:28', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e35dd41-7b4c-44e2-8076-321b9b9bfc61', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.967424] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating folder: Project (f851fef407334ba2919877cfd4865435). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.968487] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1a773e82-e874-439d-93cf-aa34a983a386 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.979439] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created folder: Project (f851fef407334ba2919877cfd4865435) in parent group-v268168. [ 1039.979647] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating folder: Instances. Parent ref: group-v268318. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1039.979890] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-09f86422-228f-45c7-a4d3-c8e68106ebf4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.988453] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created folder: Instances in parent group-v268318. [ 1039.988771] env[61907]: DEBUG oslo.service.loopingcall [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.989049] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1039.989330] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7a2529f9-4402-44f4-b30a-5f156deb228e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.014233] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.014233] env[61907]: value = "task-1244330" [ 1040.014233] env[61907]: _type = "Task" [ 1040.014233] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.023599] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244330, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.094487] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1040.125326] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.125633] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.125802] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.125993] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.126168] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.126366] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.126650] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.126842] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.127041] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.127255] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.127558] env[61907]: DEBUG nova.virt.hardware [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.128613] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49703310-dfae-4d9d-8030-76e08af56671 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.140016] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e46d4a92-ea7c-4439-aa0e-57c9b03a6023 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.175198] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244326, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.292715] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.217s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.295985] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.268s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.296212] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.296414] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1040.297353] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bccd782-3ffd-4684-8c51-aa98e8d7b138 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.306091] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e727006b-3fd7-472e-b5d2-f085b80b9a69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.321196] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4b1cd3-706b-4c7f-ae38-26a17dc2ddde {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.333754] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e0d00ab-23ed-40ef-be3d-4d328e5e1711 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.336496] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244327, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48523} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.337428] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] aaa04afe-15d8-4912-a51e-3d72bc70da9f/aaa04afe-15d8-4912-a51e-3d72bc70da9f.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1040.337677] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1040.338210] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c80d4439-97b9-4b54-a396-25ca22ecf822 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.366636] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180260MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1040.366636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.366636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.370587] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1040.370587] env[61907]: value = "task-1244331" [ 1040.370587] env[61907]: _type = "Task" [ 1040.370587] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.378949] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244331, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.415180] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Successfully updated port: b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.524823] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244330, 'name': CreateVM_Task, 'duration_secs': 0.464617} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.524988] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1040.525692] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.525873] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.526226] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.526555] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73b34249-276a-476e-92e0-4f69e8980aed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.531235] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1040.531235] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d7363c-3e20-8700-8306-f4b5b479a016" [ 1040.531235] env[61907]: _type = "Task" [ 1040.531235] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.539180] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d7363c-3e20-8700-8306-f4b5b479a016, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.672933] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244326, 'name': PowerOnVM_Task, 'duration_secs': 0.608066} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.673219] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.673430] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Took 7.26 seconds to spawn the instance on the hypervisor. [ 1040.673612] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1040.674422] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3445d079-b5b8-4a6d-8778-4de73bd2d23d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.729056] env[61907]: DEBUG nova.network.neutron [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updated VIF entry in instance network info cache for port 6e35dd41-7b4c-44e2-8076-321b9b9bfc61. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1040.729427] env[61907]: DEBUG nova.network.neutron [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.855803] env[61907]: INFO nova.scheduler.client.report [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted allocation for migration d4310bd1-9dd8-406e-a220-f078e7476274 [ 1040.883702] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244331, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086872} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.884618] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.885491] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14cd876d-aea3-405f-ac5e-00e44da224e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.909734] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] aaa04afe-15d8-4912-a51e-3d72bc70da9f/aaa04afe-15d8-4912-a51e-3d72bc70da9f.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.910367] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5b4d2ef-13d8-46f5-b9a1-d8e1705dc604 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.925505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.925505] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.925505] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1040.932557] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1040.932557] env[61907]: value = "task-1244332" [ 1040.932557] env[61907]: _type = "Task" [ 1040.932557] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.941861] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244332, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.044012] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52d7363c-3e20-8700-8306-f4b5b479a016, 'name': SearchDatastore_Task, 'duration_secs': 0.009263} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.044423] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.044626] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.044867] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.045039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.045235] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.045529] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-37e98023-5b41-4da9-a7e1-2f160dcaf0bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.053746] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.053944] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1041.054753] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1de8c21f-c127-4aad-b673-541388ca3c38 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.059937] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1041.059937] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52978789-cb88-6f0f-3d38-1860104d85a1" [ 1041.059937] env[61907]: _type = "Task" [ 1041.059937] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.068359] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52978789-cb88-6f0f-3d38-1860104d85a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.192432] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Took 14.27 seconds to build instance. [ 1041.232370] env[61907]: DEBUG oslo_concurrency.lockutils [req-8acf3df3-6dba-4cd0-aa63-75792e0b1c81 req-e7a79d49-606d-473d-bdf3-cee770cad7aa service nova] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.366039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-fbf0d623-15ab-4fd2-871d-0f7e66d15edb tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 11.525s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.395569] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.395743] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.395869] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.395990] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 3aeeac7d-6642-450e-9c1a-e43d514389a3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.396125] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 03053d82-188d-45f1-9123-240c50d25ea2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.396238] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance aaa04afe-15d8-4912-a51e-3d72bc70da9f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.396350] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f7959e23-8879-4285-93da-804dd3115f65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.396564] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance b2c5689e-f930-4548-b061-fd2974d44575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.396767] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1041.396905] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1041.446198] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244332, 'name': ReconfigVM_Task, 'duration_secs': 0.279956} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.446198] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Reconfigured VM instance instance-00000062 to attach disk [datastore2] aaa04afe-15d8-4912-a51e-3d72bc70da9f/aaa04afe-15d8-4912-a51e-3d72bc70da9f.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1041.447433] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a105510-d186-4445-8303-83fb3a6dd235 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.455070] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1041.455070] env[61907]: value = "task-1244333" [ 1041.455070] env[61907]: _type = "Task" [ 1041.455070] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.463610] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244333, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.464396] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1041.512426] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d9adaa-82f4-4df0-abef-ea6452e72e35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.522160] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949631e5-e95c-4d81-8e0d-df4ec9a72f9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.558118] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9424690a-32cb-4c6a-8db1-deeaa89b895a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.567505] env[61907]: DEBUG nova.compute.manager [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Received event network-vif-plugged-b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1041.567723] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Acquiring lock "b2c5689e-f930-4548-b061-fd2974d44575-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.567927] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Lock "b2c5689e-f930-4548-b061-fd2974d44575-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.568127] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Lock "b2c5689e-f930-4548-b061-fd2974d44575-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.568306] env[61907]: DEBUG nova.compute.manager [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] No waiting events found dispatching network-vif-plugged-b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1041.568480] env[61907]: WARNING nova.compute.manager [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Received unexpected event network-vif-plugged-b7665409-68a9-43d3-a59f-9a2562dc8923 for instance with vm_state building and task_state spawning. [ 1041.568651] env[61907]: DEBUG nova.compute.manager [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Received event network-changed-b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1041.568810] env[61907]: DEBUG nova.compute.manager [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Refreshing instance network info cache due to event network-changed-b7665409-68a9-43d3-a59f-9a2562dc8923. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1041.568978] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Acquiring lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.572839] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ec794b-6594-4327-aaa9-50b4cc7346c1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.580517] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52978789-cb88-6f0f-3d38-1860104d85a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009275} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.581595] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7962895-1ca1-4cb4-aaf6-7081949c5878 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.591569] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.595905] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1041.595905] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527cc66a-9a8a-f9d2-17f9-37b828d78bfe" [ 1041.595905] env[61907]: _type = "Task" [ 1041.595905] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.605156] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527cc66a-9a8a-f9d2-17f9-37b828d78bfe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.609337] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.609591] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.609805] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.609994] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.610183] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.612208] env[61907]: INFO nova.compute.manager [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Terminating instance [ 1041.658794] env[61907]: DEBUG nova.network.neutron [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Updating instance_info_cache with network_info: [{"id": "b7665409-68a9-43d3-a59f-9a2562dc8923", "address": "fa:16:3e:eb:4c:01", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7665409-68", "ovs_interfaceid": "b7665409-68a9-43d3-a59f-9a2562dc8923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.694424] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.780s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.965339] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244333, 'name': Rename_Task, 'duration_secs': 0.229135} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.965615] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.965911] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-abde4977-ba83-4a25-9eef-37a7d3001f7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.972295] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1041.972295] env[61907]: value = "task-1244334" [ 1041.972295] env[61907]: _type = "Task" [ 1041.972295] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.979344] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244334, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.095060] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1042.108573] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]527cc66a-9a8a-f9d2-17f9-37b828d78bfe, 'name': SearchDatastore_Task, 'duration_secs': 0.010123} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.108855] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.109159] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1042.109437] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6343f38d-699f-46cd-ae2c-8b3aa70e516c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.115290] env[61907]: DEBUG nova.compute.manager [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1042.115500] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1042.117988] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc37f4-cfb3-460e-a917-2fbb310b9463 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.121422] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1042.121422] env[61907]: value = "task-1244335" [ 1042.121422] env[61907]: _type = "Task" [ 1042.121422] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.127297] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1042.127875] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44ae0bc1-ae77-4a90-bfa4-8467983a7764 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.132562] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244335, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.137160] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1042.137160] env[61907]: value = "task-1244336" [ 1042.137160] env[61907]: _type = "Task" [ 1042.137160] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.144571] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244336, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.161763] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.162143] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Instance network_info: |[{"id": "b7665409-68a9-43d3-a59f-9a2562dc8923", "address": "fa:16:3e:eb:4c:01", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7665409-68", "ovs_interfaceid": "b7665409-68a9-43d3-a59f-9a2562dc8923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1042.162486] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Acquired lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.162723] env[61907]: DEBUG nova.network.neutron [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Refreshing network info cache for port b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1042.164008] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:4c:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b7665409-68a9-43d3-a59f-9a2562dc8923', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.172289] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating folder: Project (7581e815d75647c58ebe4198b7ed5c31). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1042.175830] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3daffc12-e200-422d-b84c-d9f7eab2f8be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.188420] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created folder: Project (7581e815d75647c58ebe4198b7ed5c31) in parent group-v268168. [ 1042.188749] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating folder: Instances. Parent ref: group-v268321. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1042.189120] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7b3f0d67-c50d-4260-8053-80961f6340e5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.199723] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created folder: Instances in parent group-v268321. [ 1042.199978] env[61907]: DEBUG oslo.service.loopingcall [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.200213] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.200434] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b750b573-7c2b-4ed1-8189-2f58e85f12d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.226560] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.226560] env[61907]: value = "task-1244339" [ 1042.226560] env[61907]: _type = "Task" [ 1042.226560] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.235052] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244339, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.482547] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244334, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.491403] env[61907]: DEBUG nova.network.neutron [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Updated VIF entry in instance network info cache for port b7665409-68a9-43d3-a59f-9a2562dc8923. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1042.491866] env[61907]: DEBUG nova.network.neutron [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Updating instance_info_cache with network_info: [{"id": "b7665409-68a9-43d3-a59f-9a2562dc8923", "address": "fa:16:3e:eb:4c:01", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb7665409-68", "ovs_interfaceid": "b7665409-68a9-43d3-a59f-9a2562dc8923", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.603277] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1042.603521] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.237s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.631907] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244335, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.475737} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.632211] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1042.632461] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.633109] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1381c319-7377-4768-a4fe-6dfd45181959 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.639224] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1042.639224] env[61907]: value = "task-1244340" [ 1042.639224] env[61907]: _type = "Task" [ 1042.639224] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.650509] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244336, 'name': PowerOffVM_Task, 'duration_secs': 0.270562} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.653412] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1042.653613] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1042.653841] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244340, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.654074] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4f7e32a-f831-46e5-81a9-efa20fd3031d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.732196] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1042.732554] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1042.732606] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleting the datastore file [datastore2] 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1042.733268] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32f5d5af-cff4-4ff3-bc2a-fb38b172dfed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.738501] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244339, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.740042] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1042.740042] env[61907]: value = "task-1244342" [ 1042.740042] env[61907]: _type = "Task" [ 1042.740042] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.747760] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244342, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.982267] env[61907]: DEBUG oslo_vmware.api [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244334, 'name': PowerOnVM_Task, 'duration_secs': 0.633881} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.982500] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.982783] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Took 7.36 seconds to spawn the instance on the hypervisor. [ 1042.982974] env[61907]: DEBUG nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1042.983760] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1de5680-0a03-4f98-b223-5481296cfcb9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.994801] env[61907]: DEBUG oslo_concurrency.lockutils [req-af996efe-7568-4ee5-aeb4-af62a746ae8d req-13d78edd-0006-45d1-8737-406fd8598652 service nova] Releasing lock "refresh_cache-b2c5689e-f930-4548-b061-fd2974d44575" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.152972] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244340, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097287} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.153279] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.154036] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61b50a62-ff62-45a3-9f00-f6578f189550 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.176211] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.176592] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d16d05b-4bfc-4258-817e-f77a85838b87 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.195777] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1043.195777] env[61907]: value = "task-1244343" [ 1043.195777] env[61907]: _type = "Task" [ 1043.195777] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.203590] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244343, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.239966] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244339, 'name': CreateVM_Task, 'duration_secs': 0.58242} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.240153] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.241038] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.241245] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.241609] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.245036] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a04e461-52cd-4c81-a133-ef57563ef04b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.251726] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1043.251726] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bd1c6a-0af6-74a8-f1d6-93b8bac6e3c3" [ 1043.251726] env[61907]: _type = "Task" [ 1043.251726] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.255299] env[61907]: DEBUG oslo_vmware.api [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244342, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16652} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.258729] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.258965] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1043.259188] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1043.259395] env[61907]: INFO nova.compute.manager [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1043.259667] env[61907]: DEBUG oslo.service.loopingcall [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.259947] env[61907]: DEBUG nova.compute.manager [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1043.259996] env[61907]: DEBUG nova.network.neutron [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1043.267548] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bd1c6a-0af6-74a8-f1d6-93b8bac6e3c3, 'name': SearchDatastore_Task, 'duration_secs': 0.008899} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.267875] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.268166] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.268443] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.268614] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.268825] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.269125] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15954d4e-86da-4aeb-ba6e-2f1fbe7357cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.277309] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.277309] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.277975] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e86540b-41a4-438b-8183-56866218842e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.283460] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1043.283460] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52754d14-031f-202d-c124-b86e636481c7" [ 1043.283460] env[61907]: _type = "Task" [ 1043.283460] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.291697] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52754d14-031f-202d-c124-b86e636481c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.499495] env[61907]: INFO nova.compute.manager [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Took 16.54 seconds to build instance. [ 1043.708191] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244343, 'name': ReconfigVM_Task, 'duration_secs': 0.266279} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.708191] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfigured VM instance instance-00000063 to attach disk [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.708795] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3dec3a4f-ac04-4538-99be-e6b7f1e63b04 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.715327] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1043.715327] env[61907]: value = "task-1244344" [ 1043.715327] env[61907]: _type = "Task" [ 1043.715327] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.724998] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244344, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.728825] env[61907]: DEBUG nova.compute.manager [req-b185412c-5f72-473d-aab3-5af8f44f25fb req-2af2831f-428c-4882-854c-210ac2d6513d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Received event network-vif-deleted-f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1043.729216] env[61907]: INFO nova.compute.manager [req-b185412c-5f72-473d-aab3-5af8f44f25fb req-2af2831f-428c-4882-854c-210ac2d6513d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Neutron deleted interface f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4; detaching it from the instance and deleting it from the info cache [ 1043.729341] env[61907]: DEBUG nova.network.neutron [req-b185412c-5f72-473d-aab3-5af8f44f25fb req-2af2831f-428c-4882-854c-210ac2d6513d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.794478] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52754d14-031f-202d-c124-b86e636481c7, 'name': SearchDatastore_Task, 'duration_secs': 0.008118} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.795721] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc7a4112-f748-421f-91e2-5bb370198753 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.801251] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1043.801251] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5202ac4a-103d-9dcb-b595-582bbf30d4db" [ 1043.801251] env[61907]: _type = "Task" [ 1043.801251] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.810485] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5202ac4a-103d-9dcb-b595-582bbf30d4db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.001526] env[61907]: DEBUG oslo_concurrency.lockutils [None req-61bffd31-902f-4b6f-aa6f-2a23f63d5c64 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.050s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.210073] env[61907]: DEBUG nova.network.neutron [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.225669] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244344, 'name': Rename_Task, 'duration_secs': 0.144986} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.225959] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1044.226228] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ddac96a7-8df5-4e9e-af41-28d0a8b6c4da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.231889] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4027fd1e-06a6-4fc5-959c-234b30cff782 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.235031] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1044.235031] env[61907]: value = "task-1244345" [ 1044.235031] env[61907]: _type = "Task" [ 1044.235031] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.241848] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e293eaed-da0a-4ab8-bf6e-33688086f354 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.255749] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244345, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.273968] env[61907]: DEBUG nova.compute.manager [req-b185412c-5f72-473d-aab3-5af8f44f25fb req-2af2831f-428c-4882-854c-210ac2d6513d service nova] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Detach interface failed, port_id=f5ac28f2-22cf-4fca-b86a-02ef8ce7b9e4, reason: Instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1044.312444] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5202ac4a-103d-9dcb-b595-582bbf30d4db, 'name': SearchDatastore_Task, 'duration_secs': 0.009077} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.312730] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.312994] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] b2c5689e-f930-4548-b061-fd2974d44575/b2c5689e-f930-4548-b061-fd2974d44575.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.313290] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4274375-f175-44c1-8767-45680e3335c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.320934] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1044.320934] env[61907]: value = "task-1244346" [ 1044.320934] env[61907]: _type = "Task" [ 1044.320934] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.329198] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244346, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.712310] env[61907]: INFO nova.compute.manager [-] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Took 1.45 seconds to deallocate network for instance. [ 1044.750426] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244345, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.796910] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.797375] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.797698] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.798016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.798290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.800626] env[61907]: INFO nova.compute.manager [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Terminating instance [ 1044.830692] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244346, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438581} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.830997] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] b2c5689e-f930-4548-b061-fd2974d44575/b2c5689e-f930-4548-b061-fd2974d44575.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.831214] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.831463] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fbc746bf-ad0f-4990-b38a-e50b87b2e8a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.838245] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1044.838245] env[61907]: value = "task-1244347" [ 1044.838245] env[61907]: _type = "Task" [ 1044.838245] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.846033] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244347, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.219025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.219313] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.219551] env[61907]: DEBUG nova.objects.instance [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'resources' on Instance uuid 92f27804-8974-40c4-9663-b2b72f0bb8e0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.247103] env[61907]: DEBUG oslo_vmware.api [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244345, 'name': PowerOnVM_Task, 'duration_secs': 0.683751} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.247441] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1045.247677] env[61907]: INFO nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 7.39 seconds to spawn the instance on the hypervisor. [ 1045.247863] env[61907]: DEBUG nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1045.248666] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1657711c-36c4-41e5-a809-bdea8513af20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.304797] env[61907]: DEBUG nova.compute.manager [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1045.305039] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1045.305980] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55057e4c-40a5-4a61-836a-facfc4295f96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.313329] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1045.313571] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-783c3955-7eac-4522-bc5b-fc62daa3ff12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.319678] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1045.319678] env[61907]: value = "task-1244348" [ 1045.319678] env[61907]: _type = "Task" [ 1045.319678] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.327553] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.347094] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244347, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079462} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.347393] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.348216] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec043fcc-c9d5-4829-9c0b-f64943dabdd0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.373661] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] b2c5689e-f930-4548-b061-fd2974d44575/b2c5689e-f930-4548-b061-fd2974d44575.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.373978] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e28ff2b8-12eb-4baf-bc1b-b98817adb5aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.393418] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1045.393418] env[61907]: value = "task-1244349" [ 1045.393418] env[61907]: _type = "Task" [ 1045.393418] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.401380] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244349, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.595454] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1045.595646] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1045.595794] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1045.767107] env[61907]: INFO nova.compute.manager [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 17.04 seconds to build instance. [ 1045.830380] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244348, 'name': PowerOffVM_Task, 'duration_secs': 0.500662} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.830670] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1045.830851] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1045.831102] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d98da318-df6a-43e2-9d5b-a9545e73f8d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.839545] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23d0c285-8cc5-44a0-a5ea-5b53c6d880e5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.846629] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32a39538-b4e0-4603-8df5-7d648cf92530 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.875802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671ca59f-e2d0-4ed0-81a0-c75b23c7c6f2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.883994] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5718dc-cd15-45f5-a084-16d0ce9fcd47 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.897158] env[61907]: DEBUG nova.compute.provider_tree [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.907018] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244349, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.013721] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1046.013980] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1046.014186] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleting the datastore file [datastore2] 3aeeac7d-6642-450e-9c1a-e43d514389a3 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.014628] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ed15681-d0ae-41dd-8705-a16883bdbc5b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.023619] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1046.023619] env[61907]: value = "task-1244351" [ 1046.023619] env[61907]: _type = "Task" [ 1046.023619] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.031105] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244351, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.272336] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26705332-a1cc-4108-8816-c0f9837687c0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.554s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.402895] env[61907]: DEBUG nova.scheduler.client.report [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1046.413486] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244349, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.459499] env[61907]: DEBUG nova.compute.manager [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Received event network-changed-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1046.459744] env[61907]: DEBUG nova.compute.manager [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Refreshing instance network info cache due to event network-changed-6e35dd41-7b4c-44e2-8076-321b9b9bfc61. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1046.459969] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.460134] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.460300] env[61907]: DEBUG nova.network.neutron [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Refreshing network info cache for port 6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1046.537222] env[61907]: DEBUG oslo_vmware.api [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244351, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.40667} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.537543] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.537750] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1046.538130] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1046.538352] env[61907]: INFO nova.compute.manager [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Took 1.23 seconds to destroy the instance on the hypervisor. [ 1046.538603] env[61907]: DEBUG oslo.service.loopingcall [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.538801] env[61907]: DEBUG nova.compute.manager [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1046.538895] env[61907]: DEBUG nova.network.neutron [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1046.907840] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244349, 'name': ReconfigVM_Task, 'duration_secs': 1.077201} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.908204] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Reconfigured VM instance instance-00000064 to attach disk [datastore1] b2c5689e-f930-4548-b061-fd2974d44575/b2c5689e-f930-4548-b061-fd2974d44575.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.909297] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.690s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.911267] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15415700-1a7c-49d3-a5c5-0d6e44a71472 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.918728] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1046.918728] env[61907]: value = "task-1244352" [ 1046.918728] env[61907]: _type = "Task" [ 1046.918728] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.928958] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244352, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.935716] env[61907]: INFO nova.scheduler.client.report [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted allocations for instance 92f27804-8974-40c4-9663-b2b72f0bb8e0 [ 1047.284590] env[61907]: DEBUG nova.network.neutron [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updated VIF entry in instance network info cache for port 6e35dd41-7b4c-44e2-8076-321b9b9bfc61. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1047.284965] env[61907]: DEBUG nova.network.neutron [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.429608] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244352, 'name': Rename_Task, 'duration_secs': 0.15493} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.429896] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.430177] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76a0367c-1546-4b4c-8a77-5efbd75c1560 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.436189] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1047.436189] env[61907]: value = "task-1244353" [ 1047.436189] env[61907]: _type = "Task" [ 1047.436189] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.445446] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244353, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.445905] env[61907]: DEBUG oslo_concurrency.lockutils [None req-63f1339f-7676-4b4b-8f49-795c7d789e18 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "92f27804-8974-40c4-9663-b2b72f0bb8e0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.836s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.648462] env[61907]: DEBUG nova.network.neutron [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.788463] env[61907]: DEBUG oslo_concurrency.lockutils [req-f1c9fd39-5743-4345-9db4-054a57302b11 req-251e1eed-651c-4fc3-a520-ef7e80a7e725 service nova] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.948468] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244353, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.151674] env[61907]: INFO nova.compute.manager [-] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Took 1.61 seconds to deallocate network for instance. [ 1048.447219] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244353, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.487731] env[61907]: DEBUG nova.compute.manager [req-62e7ebba-2ac4-4936-91de-0144bc0bea79 req-6f53b2f0-4904-46c2-ad60-dde07f9d4394 service nova] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Received event network-vif-deleted-60a000f9-0086-46cb-bf77-ed7e0b552d3b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1048.658260] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.658582] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.658934] env[61907]: DEBUG nova.objects.instance [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lazy-loading 'resources' on Instance uuid 3aeeac7d-6642-450e-9c1a-e43d514389a3 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.818568] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.818970] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.949073] env[61907]: DEBUG oslo_vmware.api [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244353, 'name': PowerOnVM_Task, 'duration_secs': 1.412361} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.949449] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.949754] env[61907]: INFO nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Took 8.85 seconds to spawn the instance on the hypervisor. [ 1048.949992] env[61907]: DEBUG nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1048.950986] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ed1625-10ef-427c-895e-2a8d3c513e24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.264687] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e32a5ae-48aa-48b3-913c-db1feacb245d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.272010] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6818d9-9535-47fc-a8e0-a62255bbcf66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.303941] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58447711-ecd0-4ff7-b32f-628e0ca1a600 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.311573] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5b7201c-737e-452c-a2e1-88fe1866386f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.325037] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1049.328432] env[61907]: DEBUG nova.compute.provider_tree [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.467463] env[61907]: INFO nova.compute.manager [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Took 18.49 seconds to build instance. [ 1049.831952] env[61907]: DEBUG nova.scheduler.client.report [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1049.846401] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.969721] env[61907]: DEBUG oslo_concurrency.lockutils [None req-24a54291-d618-4961-802b-529067a1d563 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.338676] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.341718] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.495s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.344409] env[61907]: INFO nova.compute.claims [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1050.358780] env[61907]: INFO nova.scheduler.client.report [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted allocations for instance 3aeeac7d-6642-450e-9c1a-e43d514389a3 [ 1050.629070] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1050.629255] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.866227] env[61907]: DEBUG oslo_concurrency.lockutils [None req-32d92482-220c-4e00-ab19-03f61fc80f6f tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "3aeeac7d-6642-450e-9c1a-e43d514389a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.069s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.131628] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1051.481292] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f7c0309-da5e-415a-83a7-d7bdb3f054e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.490147] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd20829-ea78-4975-b37f-64f44536ffa2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.522765] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2250bd-94d4-4c63-b99d-bba69b03c3a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.530672] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6123ae7a-eecb-4fc8-9026-16ce07426c90 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.544324] env[61907]: DEBUG nova.compute.provider_tree [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.650541] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.047691] env[61907]: DEBUG nova.scheduler.client.report [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1052.058701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "03053d82-188d-45f1-9123-240c50d25ea2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.058958] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.059215] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "03053d82-188d-45f1-9123-240c50d25ea2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.059406] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.059579] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.062064] env[61907]: INFO nova.compute.manager [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Terminating instance [ 1052.126908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.127214] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.127455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.127654] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.127831] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.129852] env[61907]: INFO nova.compute.manager [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Terminating instance [ 1052.458776] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.459087] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.552920] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.211s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.553534] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1052.556266] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.906s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.557808] env[61907]: INFO nova.compute.claims [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.565090] env[61907]: DEBUG nova.compute.manager [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1052.565300] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.566181] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3c4621-2030-4a8e-a2c0-34da3311eeb2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.573591] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.574887] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f7e4c4a-5660-43aa-8778-a4c672976aa2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.581774] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1052.581774] env[61907]: value = "task-1244354" [ 1052.581774] env[61907]: _type = "Task" [ 1052.581774] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.592031] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244354, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.633842] env[61907]: DEBUG nova.compute.manager [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1052.634048] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1052.634970] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-973e4710-4cc0-48d9-b7f1-63066bbf0a3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.642868] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1052.643140] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-838adff6-ea9a-4657-804b-491c8cf44a9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.649259] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1052.649259] env[61907]: value = "task-1244355" [ 1052.649259] env[61907]: _type = "Task" [ 1052.649259] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.657225] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.962199] env[61907]: DEBUG nova.compute.utils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.062261] env[61907]: DEBUG nova.compute.utils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1053.066438] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1053.066699] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1053.092221] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244354, 'name': PowerOffVM_Task, 'duration_secs': 0.235821} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.092570] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.092814] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.093250] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4bd1cd10-df4c-4072-a2be-bcf506a55700 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.106540] env[61907]: DEBUG nova.policy [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88b29a917e03445281f16fba6d536280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7652e98cde994af28b7bac0b81547474', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1053.161743] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244355, 'name': PowerOffVM_Task, 'duration_secs': 0.180941} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.162106] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1053.162328] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1053.162621] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d37a3077-583e-4da8-8f96-2bb87abc492f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.168972] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.169309] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.169452] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleting the datastore file [datastore2] 03053d82-188d-45f1-9123-240c50d25ea2 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.169719] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a9a5e9bb-6fe1-4d4c-a9be-63d7be626eb0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.175737] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1053.175737] env[61907]: value = "task-1244358" [ 1053.175737] env[61907]: _type = "Task" [ 1053.175737] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.182999] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244358, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.241754] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1053.242039] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1053.242255] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleting the datastore file [datastore2] aaa04afe-15d8-4912-a51e-3d72bc70da9f {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1053.242560] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d152b19-bb6b-4636-9d23-02912b20dbe5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.249097] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for the task: (returnval){ [ 1053.249097] env[61907]: value = "task-1244359" [ 1053.249097] env[61907]: _type = "Task" [ 1053.249097] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.257191] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244359, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.390769] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Successfully created port: 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1053.465759] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.567104] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1053.687209] env[61907]: DEBUG oslo_vmware.api [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244358, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267999} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.688333] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.688534] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.688741] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.688949] env[61907]: INFO nova.compute.manager [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1053.689197] env[61907]: DEBUG oslo.service.loopingcall [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.689884] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0d1e04-ab72-4cf8-a60a-e6ef18705872 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.692253] env[61907]: DEBUG nova.compute.manager [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1053.692352] env[61907]: DEBUG nova.network.neutron [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.697942] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50cee8f-ed0c-4886-a25b-c89ad7396ea1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.729611] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-061f8d6f-68fc-4d67-a5e3-7c6e574f03aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.737136] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e09ac84-aa83-4b99-a946-9fbf015e4498 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.750736] env[61907]: DEBUG nova.compute.provider_tree [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.759664] env[61907]: DEBUG oslo_vmware.api [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Task: {'id': task-1244359, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322849} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.759903] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1053.760107] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1053.760295] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1053.760474] env[61907]: INFO nova.compute.manager [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1053.760766] env[61907]: DEBUG oslo.service.loopingcall [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1053.761592] env[61907]: DEBUG nova.compute.manager [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1053.761592] env[61907]: DEBUG nova.network.neutron [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1053.928995] env[61907]: DEBUG nova.compute.manager [req-07db33ae-0a81-4e22-b5cf-068ff3e541c7 req-ae3bf11f-1705-422d-9cf8-2ca419ba0436 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Received event network-vif-deleted-c5964c65-72ff-4a67-a68a-e746390660c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1053.929265] env[61907]: INFO nova.compute.manager [req-07db33ae-0a81-4e22-b5cf-068ff3e541c7 req-ae3bf11f-1705-422d-9cf8-2ca419ba0436 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Neutron deleted interface c5964c65-72ff-4a67-a68a-e746390660c4; detaching it from the instance and deleting it from the info cache [ 1053.929504] env[61907]: DEBUG nova.network.neutron [req-07db33ae-0a81-4e22-b5cf-068ff3e541c7 req-ae3bf11f-1705-422d-9cf8-2ca419ba0436 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.997205] env[61907]: DEBUG nova.compute.manager [req-560e2555-20a7-49d9-b202-13b7d714eca4 req-1c4aa0cc-3bda-4692-9d0e-561ea7a0c74a service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Received event network-vif-deleted-0e9518d1-76b7-4257-829c-e21a49a48015 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1053.997205] env[61907]: INFO nova.compute.manager [req-560e2555-20a7-49d9-b202-13b7d714eca4 req-1c4aa0cc-3bda-4692-9d0e-561ea7a0c74a service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Neutron deleted interface 0e9518d1-76b7-4257-829c-e21a49a48015; detaching it from the instance and deleting it from the info cache [ 1053.997205] env[61907]: DEBUG nova.network.neutron [req-560e2555-20a7-49d9-b202-13b7d714eca4 req-1c4aa0cc-3bda-4692-9d0e-561ea7a0c74a service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.256052] env[61907]: DEBUG nova.scheduler.client.report [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1054.401231] env[61907]: DEBUG nova.network.neutron [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.432555] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c28a14c3-34a3-4571-9f0e-e005396b40f8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.442362] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49de2e61-46c9-4dd1-b112-092435f91f85 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.467165] env[61907]: DEBUG nova.compute.manager [req-07db33ae-0a81-4e22-b5cf-068ff3e541c7 req-ae3bf11f-1705-422d-9cf8-2ca419ba0436 service nova] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Detach interface failed, port_id=c5964c65-72ff-4a67-a68a-e746390660c4, reason: Instance 03053d82-188d-45f1-9123-240c50d25ea2 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1054.474688] env[61907]: DEBUG nova.network.neutron [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1054.499992] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e2356cfe-8630-4442-90ff-6dfde27e978f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.509018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8823a4b-0315-4a13-9e33-0b65735976aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.533394] env[61907]: DEBUG nova.compute.manager [req-560e2555-20a7-49d9-b202-13b7d714eca4 req-1c4aa0cc-3bda-4692-9d0e-561ea7a0c74a service nova] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Detach interface failed, port_id=0e9518d1-76b7-4257-829c-e21a49a48015, reason: Instance aaa04afe-15d8-4912-a51e-3d72bc70da9f could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1054.543302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.543548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.543785] env[61907]: INFO nova.compute.manager [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attaching volume 3f5c8e42-4677-4aef-942d-c3a7f7ec290c to /dev/sdb [ 1054.573730] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745e8f87-ce15-44bf-8be5-a6079aea2543 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.581018] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1054.583483] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78382255-7363-4795-af87-cd84dfe51cf5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.596285] env[61907]: DEBUG nova.virt.block_device [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating existing volume attachment record: c586e6cf-0e15-4801-8be5-515d92119e5f {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1054.605321] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1054.605576] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1054.605746] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1054.605938] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1054.606106] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1054.606265] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1054.606516] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1054.606692] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1054.606867] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1054.607074] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1054.607237] env[61907]: DEBUG nova.virt.hardware [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1054.608014] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bac3fe-3afe-4f81-b096-7e90b7bf407c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.614823] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a615225-2729-412e-aa99-6d06a9f8881f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.760907] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.761489] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1054.878518] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Successfully updated port: 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1054.903841] env[61907]: INFO nova.compute.manager [-] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Took 1.21 seconds to deallocate network for instance. [ 1054.977362] env[61907]: INFO nova.compute.manager [-] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Took 1.22 seconds to deallocate network for instance. [ 1055.266773] env[61907]: DEBUG nova.compute.utils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1055.268683] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1055.268922] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1055.306287] env[61907]: DEBUG nova.policy [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1055.381364] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.381635] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.381795] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1055.410843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.411141] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.411373] env[61907]: DEBUG nova.objects.instance [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lazy-loading 'resources' on Instance uuid 03053d82-188d-45f1-9123-240c50d25ea2 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.483263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.547322] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Successfully created port: 737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1055.772180] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1055.912636] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1055.955791] env[61907]: DEBUG nova.compute.manager [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Received event network-vif-plugged-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1055.955791] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Acquiring lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.956123] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.956190] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.956610] env[61907]: DEBUG nova.compute.manager [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] No waiting events found dispatching network-vif-plugged-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1055.956610] env[61907]: WARNING nova.compute.manager [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Received unexpected event network-vif-plugged-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 for instance with vm_state building and task_state spawning. [ 1055.956819] env[61907]: DEBUG nova.compute.manager [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Received event network-changed-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1055.957064] env[61907]: DEBUG nova.compute.manager [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Refreshing instance network info cache due to event network-changed-9fa8da7e-09f8-42c1-8705-6a29e9024ad5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1055.957277] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1056.026172] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91113671-2a15-4c81-b6d2-2b74a4761d0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.034054] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ea15ca-8beb-4c56-b8fb-da3aaf34a289 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.065515] env[61907]: DEBUG nova.network.neutron [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.067177] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccb3ef2-0b7f-41e8-8d5f-864c335da165 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.075816] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39ea0a0-7826-4754-a78a-fd75ed0f09e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.089063] env[61907]: DEBUG nova.compute.provider_tree [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.570940] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.571336] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Instance network_info: |[{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1056.572036] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1056.572036] env[61907]: DEBUG nova.network.neutron [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Refreshing network info cache for port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1056.573197] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:de:7f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9fa8da7e-09f8-42c1-8705-6a29e9024ad5', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1056.581525] env[61907]: DEBUG oslo.service.loopingcall [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1056.582644] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1056.582883] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fd6ef017-6219-410a-ac06-797bf6230efd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.597917] env[61907]: DEBUG nova.scheduler.client.report [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1056.607173] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1056.607173] env[61907]: value = "task-1244361" [ 1056.607173] env[61907]: _type = "Task" [ 1056.607173] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.616778] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244361, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.783245] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1056.811822] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1056.812125] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1056.812322] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1056.812603] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1056.812793] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1056.813015] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1056.813266] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1056.813436] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1056.813615] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1056.813785] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1056.813964] env[61907]: DEBUG nova.virt.hardware [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1056.814960] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742e4f6a-0615-4a9e-aff4-4b07c6fcba94 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.822796] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1393295e-7dc3-4bfe-9c19-82e3b4d04dea {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.905366] env[61907]: DEBUG nova.compute.manager [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Received event network-vif-plugged-737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1056.905611] env[61907]: DEBUG oslo_concurrency.lockutils [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] Acquiring lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.905826] env[61907]: DEBUG oslo_concurrency.lockutils [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.906006] env[61907]: DEBUG oslo_concurrency.lockutils [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.906193] env[61907]: DEBUG nova.compute.manager [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] No waiting events found dispatching network-vif-plugged-737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1056.906396] env[61907]: WARNING nova.compute.manager [req-4aec398f-77d1-4965-9030-32f9ee455b73 req-59c03fa8-c11f-41a5-a696-234df14aa0ab service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Received unexpected event network-vif-plugged-737a03df-dafd-4f57-80ce-f0eeb3274b46 for instance with vm_state building and task_state spawning. [ 1057.102429] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.691s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.104791] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.622s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.105048] env[61907]: DEBUG nova.objects.instance [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lazy-loading 'resources' on Instance uuid aaa04afe-15d8-4912-a51e-3d72bc70da9f {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1057.120452] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244361, 'name': CreateVM_Task, 'duration_secs': 0.301151} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.120609] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1057.121343] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.121396] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.121689] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1057.122609] env[61907]: INFO nova.scheduler.client.report [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted allocations for instance 03053d82-188d-45f1-9123-240c50d25ea2 [ 1057.123467] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae5633e2-4210-4e29-b0bd-d85f848ddb67 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.130336] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1057.130336] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb08f7-1185-44cb-8829-a3c175058961" [ 1057.130336] env[61907]: _type = "Task" [ 1057.130336] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.138029] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb08f7-1185-44cb-8829-a3c175058961, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.270228] env[61907]: DEBUG nova.network.neutron [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updated VIF entry in instance network info cache for port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1057.270603] env[61907]: DEBUG nova.network.neutron [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.478579] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Successfully updated port: 737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1057.630741] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ee733bbf-43ef-4548-927e-94171df12550 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "03053d82-188d-45f1-9123-240c50d25ea2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.572s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.643224] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52cb08f7-1185-44cb-8829-a3c175058961, 'name': SearchDatastore_Task, 'duration_secs': 0.009096} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.643528] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.643766] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1057.644013] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.644175] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.644395] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1057.645205] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-13e1e9e6-dee1-4328-943d-301ab1a8789c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.655355] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1057.655544] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1057.656746] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6a6499d-c832-49c7-a352-02a0a502c315 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.663386] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1057.663386] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5296dfdf-3fdd-f01d-cece-6e9111a28094" [ 1057.663386] env[61907]: _type = "Task" [ 1057.663386] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.670994] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5296dfdf-3fdd-f01d-cece-6e9111a28094, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.699672] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5510af98-8645-4427-9c11-3b60adf4f65c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.705879] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3511f66-25cc-42d0-b159-e7e8225f26cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.735331] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6d3da5-43ff-498f-bd18-6c7cb659532d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.744044] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ccdd984-6cf3-4ee9-bb4f-670b279ce7e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.759183] env[61907]: DEBUG nova.compute.provider_tree [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.773037] env[61907]: DEBUG oslo_concurrency.lockutils [req-7a2edbc7-5d79-4c85-8f0f-ad102e276b17 req-c42b8ec4-3d44-475d-8789-4900d3e1fe4d service nova] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1057.979243] env[61907]: DEBUG nova.compute.manager [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Received event network-changed-737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1057.979434] env[61907]: DEBUG nova.compute.manager [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Refreshing instance network info cache due to event network-changed-737a03df-dafd-4f57-80ce-f0eeb3274b46. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1057.979659] env[61907]: DEBUG oslo_concurrency.lockutils [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] Acquiring lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1057.979809] env[61907]: DEBUG oslo_concurrency.lockutils [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] Acquired lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1057.979973] env[61907]: DEBUG nova.network.neutron [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Refreshing network info cache for port 737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1057.981839] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1058.174152] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5296dfdf-3fdd-f01d-cece-6e9111a28094, 'name': SearchDatastore_Task, 'duration_secs': 0.007229} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.175093] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51b3b118-860a-429d-b311-006be6d56378 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.181755] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1058.181755] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525d1f88-76f9-b46e-5304-921596063b6c" [ 1058.181755] env[61907]: _type = "Task" [ 1058.181755] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.190371] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525d1f88-76f9-b46e-5304-921596063b6c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.261704] env[61907]: DEBUG nova.scheduler.client.report [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1058.518591] env[61907]: DEBUG nova.network.neutron [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1058.598449] env[61907]: DEBUG nova.network.neutron [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.691957] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525d1f88-76f9-b46e-5304-921596063b6c, 'name': SearchDatastore_Task, 'duration_secs': 0.029664} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.692321] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1058.692515] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1058.692875] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-955a980a-39d2-445b-9d02-99bbc5dde8e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.699089] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1058.699089] env[61907]: value = "task-1244363" [ 1058.699089] env[61907]: _type = "Task" [ 1058.699089] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.708437] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.767330] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.662s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.786469] env[61907]: INFO nova.scheduler.client.report [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Deleted allocations for instance aaa04afe-15d8-4912-a51e-3d72bc70da9f [ 1059.100866] env[61907]: DEBUG oslo_concurrency.lockutils [req-f42f4acf-a0c8-4884-8df2-22f407243d9a req-a289686d-e0a9-4814-86a0-09894e2a3e2d service nova] Releasing lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1059.101296] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1059.101502] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1059.141440] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Volume attach. Driver type: vmdk {{(pid=61907) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1059.141732] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268324', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'name': 'volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'serial': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1059.142880] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a96054-9cd0-420a-b5d3-be4f8bef1cb0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.159595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfdd486-99c2-410c-9da1-baa3814acfe1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.184925] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c/volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.185321] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-56f6ac8d-53ce-48b7-85b2-5334356dcd69 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.205085] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1059.205085] env[61907]: value = "task-1244364" [ 1059.205085] env[61907]: _type = "Task" [ 1059.205085] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.210358] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.418284} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.212048] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1059.212048] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1059.212048] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbecc785-25e3-441f-a77b-e28a82e02865 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.215872] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244364, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.220467] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1059.220467] env[61907]: value = "task-1244365" [ 1059.220467] env[61907]: _type = "Task" [ 1059.220467] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.228784] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.294852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6f39c7f6-d9a9-442f-8190-95a3ccdbb9c2 tempest-ListServersNegativeTestJSON-1218356186 tempest-ListServersNegativeTestJSON-1218356186-project-member] Lock "aaa04afe-15d8-4912-a51e-3d72bc70da9f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.167s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.646133] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1059.715782] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244364, 'name': ReconfigVM_Task, 'duration_secs': 0.317564} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.716090] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c/volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1059.720831] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-467cd4b7-0a89-42af-aae3-3f49635dedb7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.738368] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057536} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.739597] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1059.739938] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1059.739938] env[61907]: value = "task-1244366" [ 1059.739938] env[61907]: _type = "Task" [ 1059.739938] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.740627] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d70e3e-338f-4145-b504-7e549c8ada4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.769539] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1059.769861] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244366, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.770107] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-520a243f-48f1-4fa5-9b55-0fa3715c44d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.788941] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1059.788941] env[61907]: value = "task-1244367" [ 1059.788941] env[61907]: _type = "Task" [ 1059.788941] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.797682] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244367, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.872409] env[61907]: DEBUG nova.network.neutron [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Updating instance_info_cache with network_info: [{"id": "737a03df-dafd-4f57-80ce-f0eeb3274b46", "address": "fa:16:3e:5f:66:81", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap737a03df-da", "ovs_interfaceid": "737a03df-dafd-4f57-80ce-f0eeb3274b46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.253677] env[61907]: DEBUG oslo_vmware.api [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244366, 'name': ReconfigVM_Task, 'duration_secs': 0.131834} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.254030] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268324', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'name': 'volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'serial': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1060.299390] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244367, 'name': ReconfigVM_Task, 'duration_secs': 0.26026} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.299708] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.300367] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b6452ea-4d0e-445a-94e9-74fa8e2eeb4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.307020] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1060.307020] env[61907]: value = "task-1244368" [ 1060.307020] env[61907]: _type = "Task" [ 1060.307020] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.316977] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244368, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.375325] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-6d8e0c37-df00-415e-8235-6bfb9a3c3450" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1060.375680] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance network_info: |[{"id": "737a03df-dafd-4f57-80ce-f0eeb3274b46", "address": "fa:16:3e:5f:66:81", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap737a03df-da", "ovs_interfaceid": "737a03df-dafd-4f57-80ce-f0eeb3274b46", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1060.376150] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5f:66:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '737a03df-dafd-4f57-80ce-f0eeb3274b46', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1060.383779] env[61907]: DEBUG oslo.service.loopingcall [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1060.384028] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1060.384250] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8236c0a9-6525-4ff0-91d6-5a966bf45ec0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.403998] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1060.403998] env[61907]: value = "task-1244369" [ 1060.403998] env[61907]: _type = "Task" [ 1060.403998] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.412508] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244369, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.816603] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244368, 'name': Rename_Task, 'duration_secs': 0.140692} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.817736] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1060.817736] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef1dbdfc-f7de-4024-b3f1-6a82f4335892 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.823679] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1060.823679] env[61907]: value = "task-1244370" [ 1060.823679] env[61907]: _type = "Task" [ 1060.823679] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.832473] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.915611] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244369, 'name': CreateVM_Task, 'duration_secs': 0.353204} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.915770] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.916565] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.917571] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.917571] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.917571] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fcdc758-6d4d-4951-b461-e12f35dbb0ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.921832] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1060.921832] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525cd7ca-8596-911a-14c9-409757ee7337" [ 1060.921832] env[61907]: _type = "Task" [ 1060.921832] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.929781] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525cd7ca-8596-911a-14c9-409757ee7337, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.297412] env[61907]: DEBUG nova.objects.instance [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1061.334798] env[61907]: DEBUG oslo_vmware.api [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244370, 'name': PowerOnVM_Task, 'duration_secs': 0.443103} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.335102] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1061.335316] env[61907]: INFO nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1061.335501] env[61907]: DEBUG nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1061.336286] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047ca040-582e-41c1-b26a-4e465108920b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.433535] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525cd7ca-8596-911a-14c9-409757ee7337, 'name': SearchDatastore_Task, 'duration_secs': 0.012057} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.434192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.434192] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.434489] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.434688] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.434972] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.435325] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-69cfba81-ac77-481b-9dee-1d68ce1b0326 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.443951] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.444183] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.445163] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6abb28df-68bd-40df-9fb5-0d556c3e3941 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.450809] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1061.450809] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c4fe9-4efd-f2b1-7fec-beb05f494fa5" [ 1061.450809] env[61907]: _type = "Task" [ 1061.450809] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.459582] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c4fe9-4efd-f2b1-7fec-beb05f494fa5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.804023] env[61907]: DEBUG oslo_concurrency.lockutils [None req-dfce37dc-97b8-4a2a-9429-6b6cbb78c020 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.260s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.857993] env[61907]: INFO nova.compute.manager [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Took 12.03 seconds to build instance. [ 1061.961959] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c4fe9-4efd-f2b1-7fec-beb05f494fa5, 'name': SearchDatastore_Task, 'duration_secs': 0.007823} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.963059] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4813882a-98ef-4104-abea-3c11f40dbb1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.968867] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1061.968867] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528bfc1a-5e93-1a84-77fc-d136a05ad969" [ 1061.968867] env[61907]: _type = "Task" [ 1061.968867] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.978126] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528bfc1a-5e93-1a84-77fc-d136a05ad969, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.230298] env[61907]: DEBUG nova.compute.manager [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Received event network-changed-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1062.230460] env[61907]: DEBUG nova.compute.manager [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Refreshing instance network info cache due to event network-changed-9fa8da7e-09f8-42c1-8705-6a29e9024ad5. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1062.230682] env[61907]: DEBUG oslo_concurrency.lockutils [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.230837] env[61907]: DEBUG oslo_concurrency.lockutils [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.231010] env[61907]: DEBUG nova.network.neutron [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Refreshing network info cache for port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.346852] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "043b7f97-4713-431e-b5a2-be2d00c402a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.347107] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.360325] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0b32fb68-6f5c-4e8e-9a60-0732308aa015 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.541s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.479736] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528bfc1a-5e93-1a84-77fc-d136a05ad969, 'name': SearchDatastore_Task, 'duration_secs': 0.008778} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.480037] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.480307] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6d8e0c37-df00-415e-8235-6bfb9a3c3450/6d8e0c37-df00-415e-8235-6bfb9a3c3450.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.480559] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e9e9f6d6-5d02-4b43-9a58-410a12f6865c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.487508] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1062.487508] env[61907]: value = "task-1244371" [ 1062.487508] env[61907]: _type = "Task" [ 1062.487508] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.494470] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244371, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.659857] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.660486] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.849596] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1062.984179] env[61907]: DEBUG nova.network.neutron [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updated VIF entry in instance network info cache for port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1062.984596] env[61907]: DEBUG nova.network.neutron [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.996594] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244371, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458019} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.997459] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 6d8e0c37-df00-415e-8235-6bfb9a3c3450/6d8e0c37-df00-415e-8235-6bfb9a3c3450.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1062.997692] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1062.997947] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bdc549b-e234-41a7-a320-6db86ed4c0c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.004295] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1063.004295] env[61907]: value = "task-1244372" [ 1063.004295] env[61907]: _type = "Task" [ 1063.004295] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.011711] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244372, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.164965] env[61907]: DEBUG nova.compute.utils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.371904] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.372200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.373638] env[61907]: INFO nova.compute.claims [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.487247] env[61907]: DEBUG oslo_concurrency.lockutils [req-4fc1a63c-623f-4a43-8495-d9b49b75fb56 req-f1bd2d63-f17b-4a1e-b939-322c36b4b8d9 service nova] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.513153] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244372, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062755} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.513422] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1063.514225] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e279339d-7a8a-43ea-a47b-9560e8bbfcc0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.535281] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 6d8e0c37-df00-415e-8235-6bfb9a3c3450/6d8e0c37-df00-415e-8235-6bfb9a3c3450.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1063.535517] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23d7de88-0cdb-4e12-9983-fd2654ac9576 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.554058] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1063.554058] env[61907]: value = "task-1244373" [ 1063.554058] env[61907]: _type = "Task" [ 1063.554058] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.564325] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244373, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.667608] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.065021] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244373, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.472344] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9f489ab-2b14-4765-9c3f-74127c818a2d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.480166] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb99158-7a5d-4268-966b-ef0dd662ebc3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.510851] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f3b7ee-a05a-40b8-9689-e72cc2f7270f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.517720] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bee06a0-f499-4c7a-ad52-3ed548bc3a88 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.530578] env[61907]: DEBUG nova.compute.provider_tree [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.563597] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244373, 'name': ReconfigVM_Task, 'duration_secs': 0.691713} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.563870] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 6d8e0c37-df00-415e-8235-6bfb9a3c3450/6d8e0c37-df00-415e-8235-6bfb9a3c3450.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.564496] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c586ac73-d052-49a1-b8a6-19c0897b8768 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.570500] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1064.570500] env[61907]: value = "task-1244374" [ 1064.570500] env[61907]: _type = "Task" [ 1064.570500] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.578943] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244374, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.737236] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.737630] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.737851] env[61907]: INFO nova.compute.manager [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attaching volume 3b988f73-aedc-42dc-bcc0-e42c8c94c0d7 to /dev/sdc [ 1064.768812] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6622da-25d1-459b-a520-f1b6fc886f74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.775973] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc3b3039-96c7-4a08-a110-eac7ff20b9c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.788463] env[61907]: DEBUG nova.virt.block_device [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating existing volume attachment record: cabb00ec-7331-45f7-b77c-c6baeea02f58 {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1065.034182] env[61907]: DEBUG nova.scheduler.client.report [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1065.079832] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244374, 'name': Rename_Task, 'duration_secs': 0.417841} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.080190] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.080316] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8950681a-1da4-497f-86c7-56e40eba915a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.086609] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1065.086609] env[61907]: value = "task-1244376" [ 1065.086609] env[61907]: _type = "Task" [ 1065.086609] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.093754] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244376, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.539431] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.540027] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1065.596550] env[61907]: DEBUG oslo_vmware.api [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244376, 'name': PowerOnVM_Task, 'duration_secs': 0.480338} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.596864] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.597121] env[61907]: INFO nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Took 8.81 seconds to spawn the instance on the hypervisor. [ 1065.597347] env[61907]: DEBUG nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1065.598129] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280570ee-7917-4f48-9ad4-0ed968b2b1f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.044513] env[61907]: DEBUG nova.compute.utils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1066.045978] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Not allocating networking since 'none' was specified. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 1066.113328] env[61907]: INFO nova.compute.manager [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Took 14.48 seconds to build instance. [ 1066.547179] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1066.615021] env[61907]: DEBUG oslo_concurrency.lockutils [None req-aee491dc-2247-465a-875e-bc7a1576c936 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.986s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.370604] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.370986] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.371111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.371309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.371486] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.373559] env[61907]: INFO nova.compute.manager [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Terminating instance [ 1067.556237] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1067.580324] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.580583] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.580749] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.580940] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.581101] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.581263] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.581477] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.581644] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.581817] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.581985] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.582189] env[61907]: DEBUG nova.virt.hardware [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.583068] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4424ea4-f119-4ac5-8aec-d4b1d565b2a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.590802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8364f3b-88a0-45aa-b6ea-c72bd96f859d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.604143] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.609568] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Creating folder: Project (78f77d7d439746469c641b202dbcc065). Parent ref: group-v268168. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.609849] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e89e4867-05ab-434a-b38f-d86c245b3f3c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.619585] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Created folder: Project (78f77d7d439746469c641b202dbcc065) in parent group-v268168. [ 1067.619761] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Creating folder: Instances. Parent ref: group-v268328. {{(pid=61907) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1067.619980] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-733f1250-9769-458f-a12a-14a6d2c0487d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.628423] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Created folder: Instances in parent group-v268328. [ 1067.628656] env[61907]: DEBUG oslo.service.loopingcall [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.628837] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.629359] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc874454-16fe-40fa-a034-22d3c6181f81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.644352] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.644352] env[61907]: value = "task-1244380" [ 1067.644352] env[61907]: _type = "Task" [ 1067.644352] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.651174] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244380, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.877539] env[61907]: DEBUG nova.compute.manager [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1067.877772] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1067.878835] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a8e1a78-71b0-4ada-a454-6e63621c8c8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.886189] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1067.886489] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72459692-70f1-439a-a4c7-886ef26c8567 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.892707] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1067.892707] env[61907]: value = "task-1244381" [ 1067.892707] env[61907]: _type = "Task" [ 1067.892707] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.900331] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.154885] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244380, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.403278] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.656529] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244380, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.903887] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244381, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.156098] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244380, 'name': CreateVM_Task, 'duration_secs': 1.05589} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.156098] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1069.156489] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.156489] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.156757] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1069.157033] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5730169-405e-48fd-80e0-dea06e8edf91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.161622] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1069.161622] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce85c-b6ae-b911-9597-9cdf61c88ff6" [ 1069.161622] env[61907]: _type = "Task" [ 1069.161622] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.169303] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce85c-b6ae-b911-9597-9cdf61c88ff6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.330560] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Volume attach. Driver type: vmdk {{(pid=61907) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1069.330827] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268327', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'name': 'volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'serial': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1069.331730] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70b315a1-6190-4385-a64e-17d2079f24b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.347674] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a61dae6-5454-4210-ab39-0aafd9ef4219 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.376449] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7/volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.376731] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b5e4e40-dc21-4f88-bc7a-dbba70a9681c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.394584] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1069.394584] env[61907]: value = "task-1244382" [ 1069.394584] env[61907]: _type = "Task" [ 1069.394584] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.404525] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244381, 'name': PowerOffVM_Task, 'duration_secs': 1.160901} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.407887] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1069.407887] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1069.408028] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244382, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.408226] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c5e7448d-cdda-4362-8882-4ede1f67fbe3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.483134] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1069.483424] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1069.483648] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] 6d8e0c37-df00-415e-8235-6bfb9a3c3450 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1069.483938] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4722f9b-56d0-405a-a34c-9dd672a38d38 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.489940] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1069.489940] env[61907]: value = "task-1244384" [ 1069.489940] env[61907]: _type = "Task" [ 1069.489940] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.497645] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.671723] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ce85c-b6ae-b911-9597-9cdf61c88ff6, 'name': SearchDatastore_Task, 'duration_secs': 0.02909} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.672118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.672418] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1069.672708] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.672908] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.673161] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1069.673466] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-98daf5c1-1cb0-4f22-97f7-47a2b47a1c11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.681008] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1069.681239] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1069.681971] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ba619013-e987-4979-b842-ced104059a6b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.686934] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1069.686934] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276e034-a53f-c315-54e1-2b302df1013e" [ 1069.686934] env[61907]: _type = "Task" [ 1069.686934] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.693842] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276e034-a53f-c315-54e1-2b302df1013e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.903908] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244382, 'name': ReconfigVM_Task, 'duration_secs': 0.37251} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.904163] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfigured VM instance instance-0000005f to attach disk [datastore2] volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7/volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1069.909050] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1950eb46-2487-4c9c-ae62-98fd33619498 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.926025] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1069.926025] env[61907]: value = "task-1244385" [ 1069.926025] env[61907]: _type = "Task" [ 1069.926025] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.933372] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244385, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.998907] env[61907]: DEBUG oslo_vmware.api [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146079} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.999182] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1069.999377] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1069.999562] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1069.999742] env[61907]: INFO nova.compute.manager [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Took 2.12 seconds to destroy the instance on the hypervisor. [ 1069.999988] env[61907]: DEBUG oslo.service.loopingcall [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.000209] env[61907]: DEBUG nova.compute.manager [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1070.000312] env[61907]: DEBUG nova.network.neutron [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1070.196441] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5276e034-a53f-c315-54e1-2b302df1013e, 'name': SearchDatastore_Task, 'duration_secs': 0.007679} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.197236] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-28c3828c-ac54-476d-9992-02da4bc7b639 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.202084] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1070.202084] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225e5bb-5507-776f-93c9-9bffa92085dc" [ 1070.202084] env[61907]: _type = "Task" [ 1070.202084] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.209833] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225e5bb-5507-776f-93c9-9bffa92085dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.249924] env[61907]: DEBUG nova.compute.manager [req-58b7d0a1-4243-4df4-91a1-197bbf0bc9b1 req-b9ab6b3b-a4ae-413c-b5cd-b806338ffd90 service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Received event network-vif-deleted-737a03df-dafd-4f57-80ce-f0eeb3274b46 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1070.250170] env[61907]: INFO nova.compute.manager [req-58b7d0a1-4243-4df4-91a1-197bbf0bc9b1 req-b9ab6b3b-a4ae-413c-b5cd-b806338ffd90 service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Neutron deleted interface 737a03df-dafd-4f57-80ce-f0eeb3274b46; detaching it from the instance and deleting it from the info cache [ 1070.250436] env[61907]: DEBUG nova.network.neutron [req-58b7d0a1-4243-4df4-91a1-197bbf0bc9b1 req-b9ab6b3b-a4ae-413c-b5cd-b806338ffd90 service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.434176] env[61907]: DEBUG oslo_vmware.api [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244385, 'name': ReconfigVM_Task, 'duration_secs': 0.137982} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.434560] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268327', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'name': 'volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'serial': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1070.712347] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225e5bb-5507-776f-93c9-9bffa92085dc, 'name': SearchDatastore_Task, 'duration_secs': 0.008384} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.712637] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.712907] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1070.713196] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-642a39b8-9b4d-4735-91c4-9ed7b6f8e6a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.719887] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1070.719887] env[61907]: value = "task-1244386" [ 1070.719887] env[61907]: _type = "Task" [ 1070.719887] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.726864] env[61907]: DEBUG nova.network.neutron [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.727996] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244386, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.753497] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9df48817-c90b-4dd7-839e-174fb385178b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.762105] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb0eb81b-a210-4d95-9df9-a37caf38cac9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.789689] env[61907]: DEBUG nova.compute.manager [req-58b7d0a1-4243-4df4-91a1-197bbf0bc9b1 req-b9ab6b3b-a4ae-413c-b5cd-b806338ffd90 service nova] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Detach interface failed, port_id=737a03df-dafd-4f57-80ce-f0eeb3274b46, reason: Instance 6d8e0c37-df00-415e-8235-6bfb9a3c3450 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1071.229756] env[61907]: INFO nova.compute.manager [-] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Took 1.23 seconds to deallocate network for instance. [ 1071.230110] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244386, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.428942} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.231816] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1071.232099] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1071.234568] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c44d31d9-2d55-4eb4-b0db-68ab2ac5fe84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.241757] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1071.241757] env[61907]: value = "task-1244387" [ 1071.241757] env[61907]: _type = "Task" [ 1071.241757] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.248687] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244387, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.474049] env[61907]: DEBUG nova.objects.instance [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.739395] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.739767] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.740060] env[61907]: DEBUG nova.objects.instance [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid 6d8e0c37-df00-415e-8235-6bfb9a3c3450 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.752245] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244387, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06174} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.753122] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1071.753994] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e2ebbf-ff29-4c8b-856c-163eb3bd39cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.776061] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1071.776624] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b805888-20ec-42d3-a20f-289377ae05a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.798840] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1071.798840] env[61907]: value = "task-1244388" [ 1071.798840] env[61907]: _type = "Task" [ 1071.798840] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.807068] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244388, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.980136] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49a2dfa0-5a5f-42d6-a275-db3b776bd90b tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.242s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.308073] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244388, 'name': ReconfigVM_Task, 'duration_secs': 0.266197} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.310373] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1072.311132] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bf5084d7-136a-42db-bc43-5e5a546317be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.316972] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1072.316972] env[61907]: value = "task-1244389" [ 1072.316972] env[61907]: _type = "Task" [ 1072.316972] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.327624] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244389, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.339469] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6a5053b-bfb6-4725-bd49-3a629e337b1a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.345811] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc1d563-81b2-412e-9eb8-1b77c259f553 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.349000] env[61907]: DEBUG oslo_concurrency.lockutils [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.349244] env[61907]: DEBUG oslo_concurrency.lockutils [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.379401] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd50485-4ea6-4724-a138-d5977f6c6b0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.386882] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2f81aa-db39-4de8-9808-be59bd26d537 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.400418] env[61907]: DEBUG nova.compute.provider_tree [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.827432] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244389, 'name': Rename_Task, 'duration_secs': 0.124671} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.827801] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1072.827953] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55addee5-2fc2-4478-9baa-c653eb0cfed9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.833698] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1072.833698] env[61907]: value = "task-1244390" [ 1072.833698] env[61907]: _type = "Task" [ 1072.833698] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.840666] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244390, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.852506] env[61907]: INFO nova.compute.manager [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Detaching volume 3f5c8e42-4677-4aef-942d-c3a7f7ec290c [ 1072.883137] env[61907]: INFO nova.virt.block_device [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attempting to driver detach volume 3f5c8e42-4677-4aef-942d-c3a7f7ec290c from mountpoint /dev/sdb [ 1072.883458] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Volume detach. Driver type: vmdk {{(pid=61907) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1072.883671] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268324', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'name': 'volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'serial': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1072.884629] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f7b90c-dc70-471b-9b8b-94086fc8a98a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.910247] env[61907]: DEBUG nova.scheduler.client.report [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1072.914895] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54adb9e0-9de6-47b7-b09c-78cefe691194 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.921521] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2934caae-49ef-4d36-8c76-bc54d377e4ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.946129] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56b2bd6-4ff2-46ce-a5e2-a25f716631eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.960744] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] The volume has not been displaced from its original location: [datastore2] volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c/volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c.vmdk. No consolidation needed. {{(pid=61907) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1072.966163] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1072.966946] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bb32dea-d8b1-4f45-8572-575698aa59b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.984584] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1072.984584] env[61907]: value = "task-1244391" [ 1072.984584] env[61907]: _type = "Task" [ 1072.984584] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.992715] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244391, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.343929] env[61907]: DEBUG oslo_vmware.api [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244390, 'name': PowerOnVM_Task, 'duration_secs': 0.422125} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.344282] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.344531] env[61907]: INFO nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Took 5.79 seconds to spawn the instance on the hypervisor. [ 1073.344734] env[61907]: DEBUG nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1073.345484] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78c766e-8835-489a-8256-1a7101b50d13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.415108] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.441375] env[61907]: INFO nova.scheduler.client.report [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance 6d8e0c37-df00-415e-8235-6bfb9a3c3450 [ 1073.495184] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244391, 'name': ReconfigVM_Task, 'duration_secs': 0.212907} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.496303] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1073.501110] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-be315dd4-0929-417b-bcfe-b318cc5b3b8b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.516828] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1073.516828] env[61907]: value = "task-1244392" [ 1073.516828] env[61907]: _type = "Task" [ 1073.516828] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.524681] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244392, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.860918] env[61907]: INFO nova.compute.manager [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Took 10.51 seconds to build instance. [ 1073.949829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-65755299-7baa-4c56-bb91-51d4b590b337 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "6d8e0c37-df00-415e-8235-6bfb9a3c3450" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.579s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.026996] env[61907]: DEBUG oslo_vmware.api [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244392, 'name': ReconfigVM_Task, 'duration_secs': 0.14303} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.027326] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268324', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'name': 'volume-3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c', 'serial': '3f5c8e42-4677-4aef-942d-c3a7f7ec290c'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1074.176507] env[61907]: INFO nova.compute.manager [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Rebuilding instance [ 1074.211445] env[61907]: DEBUG nova.compute.manager [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1074.212349] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2d4d57d-b98e-4d39-b95e-a08d1574e3b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.362761] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7709a6b3-9c81-4f74-998a-d3b595873e14 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.015s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.565957] env[61907]: DEBUG nova.objects.instance [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1075.224985] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.225360] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32aa1ccf-9a3c-45a8-b34a-75f94e4ffe3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.233012] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1075.233012] env[61907]: value = "task-1244393" [ 1075.233012] env[61907]: _type = "Task" [ 1075.233012] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.240971] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244393, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.460606] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.460849] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.573952] env[61907]: DEBUG oslo_concurrency.lockutils [None req-768494f2-84ed-45f6-92e5-87ce45b11112 tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.225s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.590058] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.590315] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.743876] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244393, 'name': PowerOffVM_Task, 'duration_secs': 0.09927} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.744167] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.744435] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1075.745221] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01666daa-13c4-4a23-b255-8946b76cfeba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.751453] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1075.751678] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3bd3a226-856b-4c87-b8de-b6c2c46b0d6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.773014] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1075.773240] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1075.773429] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Deleting the datastore file [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.773666] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20d5a2a0-7161-4153-bcfc-121eb98b18c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.779337] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1075.779337] env[61907]: value = "task-1244395" [ 1075.779337] env[61907]: _type = "Task" [ 1075.779337] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.786612] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.962778] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1076.093125] env[61907]: INFO nova.compute.manager [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Detaching volume 3b988f73-aedc-42dc-bcc0-e42c8c94c0d7 [ 1076.127343] env[61907]: INFO nova.virt.block_device [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Attempting to driver detach volume 3b988f73-aedc-42dc-bcc0-e42c8c94c0d7 from mountpoint /dev/sdc [ 1076.127595] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Volume detach. Driver type: vmdk {{(pid=61907) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1076.127793] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268327', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'name': 'volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'serial': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1076.128666] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11673b1-090a-45bf-931c-b50e9f2e9a21 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.149283] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe0f89a-d075-4823-a558-5ad391d4bf38 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.155652] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a17fb9-5a27-4100-83e2-0a34150cb3e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.174643] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc63df36-177f-47df-b67f-38539835a597 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.188336] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] The volume has not been displaced from its original location: [datastore2] volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7/volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7.vmdk. No consolidation needed. {{(pid=61907) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1076.193681] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfiguring VM instance instance-0000005f to detach disk 2002 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1076.193946] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c6287bb-a34d-4f47-9fc0-30e7a7855b8c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.210463] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1076.210463] env[61907]: value = "task-1244396" [ 1076.210463] env[61907]: _type = "Task" [ 1076.210463] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.217615] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.289361] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102254} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.289719] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.289824] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1076.290015] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1076.485913] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.486209] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.487774] env[61907]: INFO nova.compute.claims [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.721062] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244396, 'name': ReconfigVM_Task, 'duration_secs': 0.233048} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.721062] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Reconfigured VM instance instance-0000005f to detach disk 2002 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1076.725320] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4aa6001-40db-45b2-81a1-b9d469730a93 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.740161] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1076.740161] env[61907]: value = "task-1244397" [ 1076.740161] env[61907]: _type = "Task" [ 1076.740161] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.748060] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244397, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.249910] env[61907]: DEBUG oslo_vmware.api [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244397, 'name': ReconfigVM_Task, 'duration_secs': 0.127359} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.250190] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268327', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'name': 'volume-3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'a84ffe50-2bd1-46f3-bbfd-e95f8dc23904', 'attached_at': '', 'detached_at': '', 'volume_id': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7', 'serial': '3b988f73-aedc-42dc-bcc0-e42c8c94c0d7'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1077.322587] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.322913] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.323023] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.323224] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.323376] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.323528] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.323757] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.323946] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.324112] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.324302] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.324499] env[61907]: DEBUG nova.virt.hardware [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.325373] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08597570-fa70-4bd4-98a7-2a98ab96eaa4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.333457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829b2aa8-26b1-40df-8f81-0d49f82dfda4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.346495] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance VIF info [] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.351942] env[61907]: DEBUG oslo.service.loopingcall [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.352193] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1077.352406] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1abe063e-c8aa-4437-a7fc-b797e563ffb1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.368054] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1077.368054] env[61907]: value = "task-1244398" [ 1077.368054] env[61907]: _type = "Task" [ 1077.368054] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.375193] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244398, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.590601] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d6d315-c390-46d4-8b19-3454a99beacd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.598579] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74c32c7c-0545-4409-8c4d-4bfab0b8dbf2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.628235] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b375e0f-2df8-41e2-8ba7-1f6742b55a9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.635149] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387b3f33-85ce-4177-9c33-585b570a3edd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.647493] env[61907]: DEBUG nova.compute.provider_tree [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.789545] env[61907]: DEBUG nova.objects.instance [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'flavor' on Instance uuid a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1077.877081] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244398, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.150720] env[61907]: DEBUG nova.scheduler.client.report [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1078.378288] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244398, 'name': CreateVM_Task, 'duration_secs': 0.554594} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.378616] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1078.378931] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.379176] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.379548] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1078.379844] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80be68e1-a4d7-4a5f-a665-2bcf441e2138 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.383925] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1078.383925] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52105b37-a8a9-4a6f-b694-b27eaf6856fe" [ 1078.383925] env[61907]: _type = "Task" [ 1078.383925] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.390886] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52105b37-a8a9-4a6f-b694-b27eaf6856fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.656109] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.656262] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1078.796265] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8290eeb8-de71-4a77-a28f-807a92ed2dff tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.206s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.894023] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52105b37-a8a9-4a6f-b694-b27eaf6856fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009333} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.894362] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.894611] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.894843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.894993] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.895195] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.895453] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9d50e18-c80b-4aba-b66e-48eef77fe046 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.902950] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.903138] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1078.903801] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-706a07e2-a79f-4d23-88f3-4ce8d56a61f3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.908267] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1078.908267] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c80973-ca26-7d32-e245-75fcb2c8882e" [ 1078.908267] env[61907]: _type = "Task" [ 1078.908267] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.915219] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c80973-ca26-7d32-e245-75fcb2c8882e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.161605] env[61907]: DEBUG nova.compute.utils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1079.162990] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1079.163682] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1079.207996] env[61907]: DEBUG nova.policy [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.419268] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c80973-ca26-7d32-e245-75fcb2c8882e, 'name': SearchDatastore_Task, 'duration_secs': 0.00748} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.420105] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-273559d3-08a3-45db-ba54-34aab8f02dac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.426030] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1079.426030] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523869fb-767c-f94e-a224-cbef5449cb93" [ 1079.426030] env[61907]: _type = "Task" [ 1079.426030] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.439174] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523869fb-767c-f94e-a224-cbef5449cb93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.466476] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Successfully created port: f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.667563] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1079.936761] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523869fb-767c-f94e-a224-cbef5449cb93, 'name': SearchDatastore_Task, 'duration_secs': 0.008517} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.937043] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.937369] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1079.937691] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d67e3bba-4050-40b7-a1f2-ab99b59df4df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.944682] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1079.944682] env[61907]: value = "task-1244399" [ 1079.944682] env[61907]: _type = "Task" [ 1079.944682] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.954153] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244399, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.970192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.970455] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.970677] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1079.970870] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1079.971060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.973235] env[61907]: INFO nova.compute.manager [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Terminating instance [ 1080.455446] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244399, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.418324} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.455694] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1080.455944] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.456071] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-32853761-ea59-4dd9-9039-08bc549e2c1f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.462046] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1080.462046] env[61907]: value = "task-1244400" [ 1080.462046] env[61907]: _type = "Task" [ 1080.462046] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.469542] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244400, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.477686] env[61907]: DEBUG nova.compute.manager [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1080.477953] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.478920] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb31b23-76b1-4b9e-9e0d-632b7f6653bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.485816] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1080.486042] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-47e17aa1-a687-429d-9fc2-935e98608e42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.491756] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1080.491756] env[61907]: value = "task-1244401" [ 1080.491756] env[61907]: _type = "Task" [ 1080.491756] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.499758] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244401, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.679585] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1080.704101] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.704549] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.704839] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.705135] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.705306] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.705466] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.705679] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.705847] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.706035] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.706209] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.706420] env[61907]: DEBUG nova.virt.hardware [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.707335] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18e11ee-fe1b-4c3b-b5e4-67df255ce137 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.715018] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f60c382-d4dc-4522-ad25-62e0cc6bf673 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.916905] env[61907]: DEBUG nova.compute.manager [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Received event network-vif-plugged-f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1080.917156] env[61907]: DEBUG oslo_concurrency.lockutils [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] Acquiring lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.917404] env[61907]: DEBUG oslo_concurrency.lockutils [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.917620] env[61907]: DEBUG oslo_concurrency.lockutils [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.917799] env[61907]: DEBUG nova.compute.manager [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] No waiting events found dispatching network-vif-plugged-f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1080.917974] env[61907]: WARNING nova.compute.manager [req-eb449a79-a163-480b-99a9-758cabbd8cbc req-0a9882f0-bf78-41c7-8318-459b3832ab7b service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Received unexpected event network-vif-plugged-f82c11cd-e184-4996-a58b-357ab9c9627c for instance with vm_state building and task_state spawning. [ 1080.974635] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244400, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06085} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.974893] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1080.976026] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56dad96f-8b5f-414b-b39e-963576a81d0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.999301] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1080.999647] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8cde070a-a8c8-4fd9-9119-80c4c28a05f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.020789] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244401, 'name': PowerOffVM_Task, 'duration_secs': 0.215757} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.021974] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1081.022183] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1081.022502] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1081.022502] env[61907]: value = "task-1244402" [ 1081.022502] env[61907]: _type = "Task" [ 1081.022502] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.022690] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7ff9e8f-cc89-4589-b3bb-f18ea2ce171e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.031450] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244402, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.101379] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1081.101716] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1081.101794] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleting the datastore file [datastore1] a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1081.102068] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ea0b7ad-9b32-4753-9bde-61cabee976c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.108652] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for the task: (returnval){ [ 1081.108652] env[61907]: value = "task-1244404" [ 1081.108652] env[61907]: _type = "Task" [ 1081.108652] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.117145] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244404, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.215665] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Successfully updated port: f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1081.532989] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244402, 'name': ReconfigVM_Task, 'duration_secs': 0.278008} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.533331] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Reconfigured VM instance instance-00000067 to attach disk [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3/043b7f97-4713-431e-b5a2-be2d00c402a3.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1081.533940] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7693135-ada9-4407-abdf-e3560625d802 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.540305] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1081.540305] env[61907]: value = "task-1244405" [ 1081.540305] env[61907]: _type = "Task" [ 1081.540305] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.547472] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244405, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.618487] env[61907]: DEBUG oslo_vmware.api [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Task: {'id': task-1244404, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172348} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.618802] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1081.618962] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1081.619169] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1081.619354] env[61907]: INFO nova.compute.manager [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1081.619603] env[61907]: DEBUG oslo.service.loopingcall [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1081.619805] env[61907]: DEBUG nova.compute.manager [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1081.619907] env[61907]: DEBUG nova.network.neutron [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.718724] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.718930] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.719024] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1082.051169] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244405, 'name': Rename_Task, 'duration_secs': 0.133761} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.051169] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1082.051169] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b2022fb5-c4f2-4a0f-ab42-c0bea5442aa7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.059061] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1082.059061] env[61907]: value = "task-1244406" [ 1082.059061] env[61907]: _type = "Task" [ 1082.059061] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.070735] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244406, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.256137] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1082.379234] env[61907]: DEBUG nova.network.neutron [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Updating instance_info_cache with network_info: [{"id": "f82c11cd-e184-4996-a58b-357ab9c9627c", "address": "fa:16:3e:cc:61:dd", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf82c11cd-e1", "ovs_interfaceid": "f82c11cd-e184-4996-a58b-357ab9c9627c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.510845] env[61907]: DEBUG nova.network.neutron [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.568640] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244406, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.882090] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.882464] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Instance network_info: |[{"id": "f82c11cd-e184-4996-a58b-357ab9c9627c", "address": "fa:16:3e:cc:61:dd", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf82c11cd-e1", "ovs_interfaceid": "f82c11cd-e184-4996-a58b-357ab9c9627c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1082.882924] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cc:61:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f82c11cd-e184-4996-a58b-357ab9c9627c', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.890445] env[61907]: DEBUG oslo.service.loopingcall [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.890662] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.890884] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5df5e567-8bd4-400b-8769-152b481ba81e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.910251] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.910251] env[61907]: value = "task-1244407" [ 1082.910251] env[61907]: _type = "Task" [ 1082.910251] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.917798] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244407, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.946139] env[61907]: DEBUG nova.compute.manager [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Received event network-changed-f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1082.946327] env[61907]: DEBUG nova.compute.manager [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Refreshing instance network info cache due to event network-changed-f82c11cd-e184-4996-a58b-357ab9c9627c. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1082.946570] env[61907]: DEBUG oslo_concurrency.lockutils [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] Acquiring lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.946723] env[61907]: DEBUG oslo_concurrency.lockutils [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] Acquired lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.946891] env[61907]: DEBUG nova.network.neutron [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Refreshing network info cache for port f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1083.013786] env[61907]: INFO nova.compute.manager [-] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Took 1.39 seconds to deallocate network for instance. [ 1083.068792] env[61907]: DEBUG oslo_vmware.api [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244406, 'name': PowerOnVM_Task, 'duration_secs': 0.513663} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.069097] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1083.069311] env[61907]: DEBUG nova.compute.manager [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1083.070070] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a826ca0-ba23-44cc-adbe-11d9e40f01e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.420410] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244407, 'name': CreateVM_Task, 'duration_secs': 0.386825} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.420577] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.421292] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.421509] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.421885] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.422398] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-485aa596-54df-4a01-b2f9-49f44d866815 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.426805] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1083.426805] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a767ae-594d-46ab-185a-326993cbe833" [ 1083.426805] env[61907]: _type = "Task" [ 1083.426805] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.434487] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a767ae-594d-46ab-185a-326993cbe833, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.522473] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.522473] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.522585] env[61907]: DEBUG nova.objects.instance [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lazy-loading 'resources' on Instance uuid a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1083.587095] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.685414] env[61907]: DEBUG nova.network.neutron [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Updated VIF entry in instance network info cache for port f82c11cd-e184-4996-a58b-357ab9c9627c. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1083.685782] env[61907]: DEBUG nova.network.neutron [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Updating instance_info_cache with network_info: [{"id": "f82c11cd-e184-4996-a58b-357ab9c9627c", "address": "fa:16:3e:cc:61:dd", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf82c11cd-e1", "ovs_interfaceid": "f82c11cd-e184-4996-a58b-357ab9c9627c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.890180] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "043b7f97-4713-431e-b5a2-be2d00c402a3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.890413] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.890627] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "043b7f97-4713-431e-b5a2-be2d00c402a3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.890812] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.890984] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.892982] env[61907]: INFO nova.compute.manager [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Terminating instance [ 1083.937451] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52a767ae-594d-46ab-185a-326993cbe833, 'name': SearchDatastore_Task, 'duration_secs': 0.012179} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.937743] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.937979] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.938229] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.938384] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.938570] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.938829] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-706a8292-8ac1-4dea-9625-169c8aa55457 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.946620] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.946805] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.947541] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da0073f4-fbfa-4287-bff7-0cc321b14e49 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.952486] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1083.952486] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5289fb73-3f21-4812-3041-20d1414b88c3" [ 1083.952486] env[61907]: _type = "Task" [ 1083.952486] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.959603] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5289fb73-3f21-4812-3041-20d1414b88c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.112362] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e8c93b-34ae-4965-980f-67373c8ed9d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.119747] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9698d497-f28b-43e8-ba2d-82c381365fa4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.148426] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b5e8fa3-6383-49f5-8f90-58538e2e726a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.155019] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf91b863-dd6a-47f5-b551-9666ee0ccca5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.167797] env[61907]: DEBUG nova.compute.provider_tree [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.188329] env[61907]: DEBUG oslo_concurrency.lockutils [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] Releasing lock "refresh_cache-32fc7dba-6a76-4f0b-bca2-204e3bc228de" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.188571] env[61907]: DEBUG nova.compute.manager [req-58763ec1-b59a-4054-9de0-3b1cad005537 req-46fd8c49-2712-4f28-998e-c045f9228ba6 service nova] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Received event network-vif-deleted-cd2253b4-0244-45b6-80cb-b483ff82a146 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1084.306930] env[61907]: DEBUG nova.compute.manager [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1084.308069] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dec054-5dfd-4dab-af41-0d85454dd1f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.397430] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "refresh_cache-043b7f97-4713-431e-b5a2-be2d00c402a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.397814] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquired lock "refresh_cache-043b7f97-4713-431e-b5a2-be2d00c402a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.398121] env[61907]: DEBUG nova.network.neutron [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1084.462530] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5289fb73-3f21-4812-3041-20d1414b88c3, 'name': SearchDatastore_Task, 'duration_secs': 0.008419} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.463322] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b14a2da1-6b06-40fc-9d88-1c8ba86bad82 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.468446] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1084.468446] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e696a8-1d68-bf9c-0016-0ea3999b8946" [ 1084.468446] env[61907]: _type = "Task" [ 1084.468446] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.476017] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e696a8-1d68-bf9c-0016-0ea3999b8946, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.671027] env[61907]: DEBUG nova.scheduler.client.report [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1084.819821] env[61907]: INFO nova.compute.manager [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] instance snapshotting [ 1084.820907] env[61907]: DEBUG nova.objects.instance [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.916118] env[61907]: DEBUG nova.network.neutron [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1084.974183] env[61907]: DEBUG nova.network.neutron [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.980669] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52e696a8-1d68-bf9c-0016-0ea3999b8946, 'name': SearchDatastore_Task, 'duration_secs': 0.009845} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.980938] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.981219] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 32fc7dba-6a76-4f0b-bca2-204e3bc228de/32fc7dba-6a76-4f0b-bca2-204e3bc228de.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.981479] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0dcf58d-729b-4ea1-a990-d8bec5c7aa62 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.988933] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1084.988933] env[61907]: value = "task-1244408" [ 1084.988933] env[61907]: _type = "Task" [ 1084.988933] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.996738] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.177918] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.656s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.180490] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 1.594s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.180732] env[61907]: DEBUG nova.objects.instance [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61907) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1085.204795] env[61907]: INFO nova.scheduler.client.report [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Deleted allocations for instance a84ffe50-2bd1-46f3-bbfd-e95f8dc23904 [ 1085.326722] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58df4d8d-7e77-4072-9e65-bb9c247b144b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.346802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0626f176-3dea-424a-bf3e-129046f610d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.477148] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Releasing lock "refresh_cache-043b7f97-4713-431e-b5a2-be2d00c402a3" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.477734] env[61907]: DEBUG nova.compute.manager [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1085.477935] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1085.478822] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd492cbc-edb7-4e31-aa65-59c156595001 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.485823] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1085.486074] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14cba0eb-7bef-4e6b-a8be-c4ce76f3f4d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.493593] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1085.493593] env[61907]: value = "task-1244409" [ 1085.493593] env[61907]: _type = "Task" [ 1085.493593] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.499472] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43824} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.499991] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 32fc7dba-6a76-4f0b-bca2-204e3bc228de/32fc7dba-6a76-4f0b-bca2-204e3bc228de.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.500225] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.500461] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f6241e22-5fee-4031-921d-fe2ef132baf0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.504583] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.508614] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1085.508614] env[61907]: value = "task-1244410" [ 1085.508614] env[61907]: _type = "Task" [ 1085.508614] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.515971] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244410, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.714092] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f0664f89-2557-49bb-b00f-8812c737d2ae tempest-AttachVolumeTestJSON-2048998921 tempest-AttachVolumeTestJSON-2048998921-project-member] Lock "a84ffe50-2bd1-46f3-bbfd-e95f8dc23904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.743s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1085.860222] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1085.861112] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ac1fa5eb-029c-479e-af71-3cfc39673601 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.868839] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1085.868839] env[61907]: value = "task-1244411" [ 1085.868839] env[61907]: _type = "Task" [ 1085.868839] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.876456] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244411, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.003466] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244409, 'name': PowerOffVM_Task, 'duration_secs': 0.108834} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.003682] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1086.003856] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1086.004122] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8119a093-fbe1-438f-a692-4e1f920c326b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.016997] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244410, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065517} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.017315] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1086.018148] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0948c7cd-6fea-4d67-b52b-7b5a9b8f4b13 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.043050] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Reconfiguring VM instance instance-00000068 to attach disk [datastore2] 32fc7dba-6a76-4f0b-bca2-204e3bc228de/32fc7dba-6a76-4f0b-bca2-204e3bc228de.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1086.044268] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0798edf-fb21-4293-a336-09e4df8a8f3f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.057993] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1086.058230] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1086.058417] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Deleting the datastore file [datastore2] 043b7f97-4713-431e-b5a2-be2d00c402a3 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1086.058699] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7feaee99-fcf5-4c2c-9988-4094abc8b08b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.065430] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for the task: (returnval){ [ 1086.065430] env[61907]: value = "task-1244413" [ 1086.065430] env[61907]: _type = "Task" [ 1086.065430] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.066686] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1086.066686] env[61907]: value = "task-1244414" [ 1086.066686] env[61907]: _type = "Task" [ 1086.066686] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.076231] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244413, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.079169] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244414, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.189830] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d70f974f-ce98-46e9-a3ac-3bd52ab2b1d9 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.379452] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244411, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.580514] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244414, 'name': ReconfigVM_Task, 'duration_secs': 0.302596} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.583684] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Reconfigured VM instance instance-00000068 to attach disk [datastore2] 32fc7dba-6a76-4f0b-bca2-204e3bc228de/32fc7dba-6a76-4f0b-bca2-204e3bc228de.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.584475] env[61907]: DEBUG oslo_vmware.api [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Task: {'id': task-1244413, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.110612} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.584687] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a40f137-2dde-441c-a676-1f22a8a1ffa0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.586197] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1086.586415] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1086.586605] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1086.586782] env[61907]: INFO nova.compute.manager [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1086.587023] env[61907]: DEBUG oslo.service.loopingcall [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1086.587229] env[61907]: DEBUG nova.compute.manager [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1086.587333] env[61907]: DEBUG nova.network.neutron [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1086.595402] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1086.595402] env[61907]: value = "task-1244415" [ 1086.595402] env[61907]: _type = "Task" [ 1086.595402] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.600658] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244415, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.603399] env[61907]: DEBUG nova.network.neutron [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1086.880150] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244411, 'name': CreateSnapshot_Task, 'duration_secs': 0.562668} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.880150] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1086.880595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13f85b9d-8183-493b-8d51-b58da07f7818 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.103420] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244415, 'name': Rename_Task, 'duration_secs': 0.134292} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.103657] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.103920] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5078c89-448d-45be-a4e0-1795e3b96684 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.110125] env[61907]: DEBUG nova.network.neutron [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1087.111257] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1087.111257] env[61907]: value = "task-1244417" [ 1087.111257] env[61907]: _type = "Task" [ 1087.111257] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.118635] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244417, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.399148] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1087.399474] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-dc0df907-9552-4c53-9185-f42b9b8c6ece {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.410682] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1087.410682] env[61907]: value = "task-1244418" [ 1087.410682] env[61907]: _type = "Task" [ 1087.410682] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.418884] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244418, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.613108] env[61907]: INFO nova.compute.manager [-] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Took 1.03 seconds to deallocate network for instance. [ 1087.626841] env[61907]: DEBUG oslo_vmware.api [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244417, 'name': PowerOnVM_Task, 'duration_secs': 0.475688} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.627222] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1087.627534] env[61907]: INFO nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Took 6.95 seconds to spawn the instance on the hypervisor. [ 1087.627798] env[61907]: DEBUG nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1087.628718] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebcd6ec-db8a-4bc0-a348-f6674ce3891e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.920173] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244418, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.123217] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.123503] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.123981] env[61907]: DEBUG nova.objects.instance [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lazy-loading 'resources' on Instance uuid 043b7f97-4713-431e-b5a2-be2d00c402a3 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1088.144926] env[61907]: INFO nova.compute.manager [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Took 11.68 seconds to build instance. [ 1088.421179] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244418, 'name': CloneVM_Task, 'duration_secs': 0.955567} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.421455] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created linked-clone VM from snapshot [ 1088.422211] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77066d34-5bdc-4594-8416-8e5d3107f6a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.429538] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploading image 4f4d5745-4c73-47d1-8075-14da12669c40 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1088.451143] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1088.451143] env[61907]: value = "vm-268334" [ 1088.451143] env[61907]: _type = "VirtualMachine" [ 1088.451143] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1088.451724] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-311d76ee-524f-4069-b8b1-bce4ff5f8e70 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.458614] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease: (returnval){ [ 1088.458614] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52136633-b027-a9a9-a219-b8c0c05b4116" [ 1088.458614] env[61907]: _type = "HttpNfcLease" [ 1088.458614] env[61907]: } obtained for exporting VM: (result){ [ 1088.458614] env[61907]: value = "vm-268334" [ 1088.458614] env[61907]: _type = "VirtualMachine" [ 1088.458614] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1088.458888] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the lease: (returnval){ [ 1088.458888] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52136633-b027-a9a9-a219-b8c0c05b4116" [ 1088.458888] env[61907]: _type = "HttpNfcLease" [ 1088.458888] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1088.465424] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1088.465424] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52136633-b027-a9a9-a219-b8c0c05b4116" [ 1088.465424] env[61907]: _type = "HttpNfcLease" [ 1088.465424] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1088.649060] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a325390e-d02a-4700-b395-2231fca2a396 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.188s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.723268] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b524f8-9cb8-4109-b4cd-2f0908ce11cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.730711] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853d0fbb-584b-4efd-a01f-cedffff0e74c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.764510] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d71706-6441-4355-b2c3-7d50c4ee8d1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.771548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7908692f-38e1-4051-b302-2ce429a8523f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.785131] env[61907]: DEBUG nova.compute.provider_tree [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.809124] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.809389] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.809610] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.809804] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.809980] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.812046] env[61907]: INFO nova.compute.manager [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Terminating instance [ 1088.969175] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1088.969175] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52136633-b027-a9a9-a219-b8c0c05b4116" [ 1088.969175] env[61907]: _type = "HttpNfcLease" [ 1088.969175] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1088.969625] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1088.969625] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52136633-b027-a9a9-a219-b8c0c05b4116" [ 1088.969625] env[61907]: _type = "HttpNfcLease" [ 1088.969625] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1088.970197] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43885b0d-f31b-4435-9936-6c8405d757b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.977899] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1088.978095] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1089.072645] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-36c5735f-8355-4822-802c-787eafc90055 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.291320] env[61907]: DEBUG nova.scheduler.client.report [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1089.316212] env[61907]: DEBUG nova.compute.manager [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1089.316675] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1089.317654] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5fd745-b256-40b3-b8ea-dfa281a29e96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.326369] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1089.326654] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e25bf30e-4482-4109-b3f9-53b50d6d1b93 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.333556] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1089.333556] env[61907]: value = "task-1244420" [ 1089.333556] env[61907]: _type = "Task" [ 1089.333556] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.342109] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244420, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.514569] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1089.514737] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances with incomplete migration {{(pid=61907) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 1089.796978] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.826659] env[61907]: INFO nova.scheduler.client.report [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Deleted allocations for instance 043b7f97-4713-431e-b5a2-be2d00c402a3 [ 1089.843621] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244420, 'name': PowerOffVM_Task, 'duration_secs': 0.19777} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.843921] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1089.844133] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1089.844555] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79f2e3df-38aa-4454-8c4e-9ac236d94f8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.905305] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1089.905741] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1089.905858] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] 32fc7dba-6a76-4f0b-bca2-204e3bc228de {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.906156] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-76ee8886-2389-4035-8b74-0e66ea85121f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.913674] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1089.913674] env[61907]: value = "task-1244423" [ 1089.913674] env[61907]: _type = "Task" [ 1089.913674] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.923293] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.336197] env[61907]: DEBUG oslo_concurrency.lockutils [None req-6708fb64-775c-4257-99de-c55a557acdf2 tempest-ServerShowV254Test-1917028345 tempest-ServerShowV254Test-1917028345-project-member] Lock "043b7f97-4713-431e-b5a2-be2d00c402a3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.446s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.424882] env[61907]: DEBUG oslo_vmware.api [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.178416} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.425323] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.425712] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1090.425985] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1090.426358] env[61907]: INFO nova.compute.manager [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1090.426680] env[61907]: DEBUG oslo.service.loopingcall [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1090.427020] env[61907]: DEBUG nova.compute.manager [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1090.427147] env[61907]: DEBUG nova.network.neutron [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1090.721549] env[61907]: DEBUG nova.compute.manager [req-2956c263-acaa-4b05-8197-fb22cf9e31d5 req-42a68633-a279-45e3-ace7-b3dde6c8d88c service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Received event network-vif-deleted-f82c11cd-e184-4996-a58b-357ab9c9627c {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1090.721809] env[61907]: INFO nova.compute.manager [req-2956c263-acaa-4b05-8197-fb22cf9e31d5 req-42a68633-a279-45e3-ace7-b3dde6c8d88c service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Neutron deleted interface f82c11cd-e184-4996-a58b-357ab9c9627c; detaching it from the instance and deleting it from the info cache [ 1090.721984] env[61907]: DEBUG nova.network.neutron [req-2956c263-acaa-4b05-8197-fb22cf9e31d5 req-42a68633-a279-45e3-ace7-b3dde6c8d88c service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.195819] env[61907]: DEBUG nova.network.neutron [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.225833] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f790c795-b23d-426d-a278-26e4a37473ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.235217] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0242e37-7862-4bbf-80c3-f8a7f03d484c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.263013] env[61907]: DEBUG nova.compute.manager [req-2956c263-acaa-4b05-8197-fb22cf9e31d5 req-42a68633-a279-45e3-ace7-b3dde6c8d88c service nova] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Detach interface failed, port_id=f82c11cd-e184-4996-a58b-357ab9c9627c, reason: Instance 32fc7dba-6a76-4f0b-bca2-204e3bc228de could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1091.698391] env[61907]: INFO nova.compute.manager [-] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Took 1.27 seconds to deallocate network for instance. [ 1092.017363] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1092.017711] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 1092.205426] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.205729] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.205962] env[61907]: DEBUG nova.objects.instance [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid 32fc7dba-6a76-4f0b-bca2-204e3bc228de {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1092.535509] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] There are 52 instances to clean {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 1092.535816] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 043b7f97-4713-431e-b5a2-be2d00c402a3] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1092.792526] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f88e81bd-5d17-41a4-98a7-73f808031e1f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.800265] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8ef51e3-d632-471d-a3f9-3e0ed2deda93 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.830563] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51d44d7f-f673-449a-b2ac-9577aeb609e5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.838146] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe5b6df1-3af8-429d-9333-80c55cd0f7fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.852953] env[61907]: DEBUG nova.compute.provider_tree [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.039442] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 6d8e0c37-df00-415e-8235-6bfb9a3c3450] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1093.356492] env[61907]: DEBUG nova.scheduler.client.report [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1093.543290] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: aaa04afe-15d8-4912-a51e-3d72bc70da9f] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1093.862141] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.656s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.881932] env[61907]: INFO nova.scheduler.client.report [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance 32fc7dba-6a76-4f0b-bca2-204e3bc228de [ 1094.047444] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 03053d82-188d-45f1-9123-240c50d25ea2] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1094.388995] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9fe6a133-fe90-4157-b021-6880e39824ea tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "32fc7dba-6a76-4f0b-bca2-204e3bc228de" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.579s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.550643] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 3aeeac7d-6642-450e-9c1a-e43d514389a3] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1095.054217] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a84ffe50-2bd1-46f3-bbfd-e95f8dc23904] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1095.557951] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 6046a82d-f73d-4369-8208-c5a2e5c77431] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1095.959650] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.959929] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.061792] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 696c8d63-5268-4239-8d41-079e9f3f7c66] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1096.203643] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1096.204773] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd24e17-0c72-4e52-8182-1bca0d688d57 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.211045] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1096.211258] env[61907]: ERROR oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk due to incomplete transfer. [ 1096.211511] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-5c8c7038-2317-4f76-bb7c-297851f380e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.219252] env[61907]: DEBUG oslo_vmware.rw_handles [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5287a30f-a218-bfc8-795e-0eb965dd947c/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1096.219459] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploaded image 4f4d5745-4c73-47d1-8075-14da12669c40 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1096.221726] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1096.222044] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-68240121-d31c-4e2b-a21b-b65f9a18b0be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.229677] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1096.229677] env[61907]: value = "task-1244425" [ 1096.229677] env[61907]: _type = "Task" [ 1096.229677] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.237194] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244425, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.461831] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1096.565164] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 4430db3b-0276-42b6-a67e-386a164ddc0e] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1096.739016] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244425, 'name': Destroy_Task, 'duration_secs': 0.329086} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.739287] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroyed the VM [ 1096.739535] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1096.739794] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-03ef3bae-e441-4fd1-b21a-cf1a2200441c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.745818] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1096.745818] env[61907]: value = "task-1244426" [ 1096.745818] env[61907]: _type = "Task" [ 1096.745818] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.753308] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244426, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.984069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.984425] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.985894] env[61907]: INFO nova.compute.claims [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1097.067972] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: b43efe48-95de-46ec-8cbb-c24cf7bd68a1] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1097.254878] env[61907]: DEBUG oslo_vmware.api [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244426, 'name': RemoveSnapshot_Task, 'duration_secs': 0.486987} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.255179] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1097.255419] env[61907]: INFO nova.compute.manager [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 11.93 seconds to snapshot the instance on the hypervisor. [ 1097.571230] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 6a092752-058f-4196-848e-58a94d084cc7] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1097.812900] env[61907]: DEBUG nova.compute.manager [None req-e25e931e-cf4e-4d56-ba40-d6d61a2f431b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Found 1 images (rotation: 2) {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 1098.064465] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb7dadb-527f-4859-ad10-9111f93bbd11 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.071874] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-711c14bd-e0e2-468d-9d63-643d2e270f54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.075421] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f50bf687-609c-44e1-bd75-cf4efaeadb71] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1098.102522] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5475c612-e718-49eb-9760-9cfedbd7931d] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1098.104953] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f00b80-fe3b-47c5-aa17-c484b608f59b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.112719] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c32a6a3-aa30-46af-a606-78137de1911b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.126647] env[61907]: DEBUG nova.compute.provider_tree [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1098.608988] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f87e45da-187b-4aad-b7bf-b4228a8b2f1f] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1098.629422] env[61907]: DEBUG nova.scheduler.client.report [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1099.112658] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: cd2c70a7-a290-4d1a-91fe-5da772860603] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1099.134935] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.150s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.135562] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1099.169063] env[61907]: DEBUG nova.compute.manager [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1099.170178] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f344e4-2fcf-41dc-9367-3eeaf2d2a529 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.615718] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: fcde07bf-858a-4377-b27c-4f17356306af] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1099.640593] env[61907]: DEBUG nova.compute.utils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.642464] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1099.642633] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1099.680017] env[61907]: INFO nova.compute.manager [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] instance snapshotting [ 1099.680593] env[61907]: DEBUG nova.objects.instance [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.683467] env[61907]: DEBUG nova.policy [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1099.940366] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Successfully created port: ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1100.119390] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 23f9f3d1-b376-4502-927c-e9bd89693131] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1100.146967] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1100.188884] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2683ac59-6cc7-404c-b172-deddfe1c96b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.207727] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbf776b1-6185-4365-8e72-837120af0557 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.623148] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 71804bc1-79ce-4731-9e27-74c007d4e906] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1100.704205] env[61907]: DEBUG nova.compute.manager [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1100.718650] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1100.718989] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-29f23fd2-da46-4f59-a06e-52fa4904f5e7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.735075] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1100.735075] env[61907]: value = "task-1244427" [ 1100.735075] env[61907]: _type = "Task" [ 1100.735075] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.744040] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244427, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.204979] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1101.207259] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c3bf9001-7166-4064-89d5-96be073fa7c2] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1101.233972] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1101.233972] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1101.234154] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1101.234266] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1101.234423] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1101.234581] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1101.234797] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1101.234963] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1101.235454] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1101.235454] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1101.235515] env[61907]: DEBUG nova.virt.hardware [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1101.236427] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28cd9872-b68f-4049-8d25-6f2bea085b8d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.252273] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7beebab-8d06-41d7-8a16-96400c0fd5a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.255904] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244427, 'name': CreateSnapshot_Task, 'duration_secs': 0.401972} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.256519] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1101.259078] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331a63cb-a8ff-4b62-9ddf-0921f1fe658d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.315843] env[61907]: DEBUG nova.compute.manager [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Received event network-vif-plugged-ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1101.316095] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] Acquiring lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.316364] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.316548] env[61907]: DEBUG oslo_concurrency.lockutils [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.316730] env[61907]: DEBUG nova.compute.manager [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] No waiting events found dispatching network-vif-plugged-ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1101.316888] env[61907]: WARNING nova.compute.manager [req-2f98c169-976a-4b47-9fcb-d9c25e924b70 req-cc349b67-c05f-4e0a-95f6-2fb016c62bb5 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Received unexpected event network-vif-plugged-ce2db7f4-56a8-43ee-904e-1e7c22aa7512 for instance with vm_state building and task_state spawning. [ 1101.323661] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.323886] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.393603] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Successfully updated port: ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1101.712869] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 4b07210f-8005-481f-b7a1-ce80a4cbe980] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1101.783493] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1101.783739] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b1494199-bf2d-4603-a346-292dbe0cc6c5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.793469] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1101.793469] env[61907]: value = "task-1244428" [ 1101.793469] env[61907]: _type = "Task" [ 1101.793469] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.801161] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244428, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.828950] env[61907]: INFO nova.compute.claims [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1101.896417] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.896646] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.896829] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.216595] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a31110ed-c4bb-4492-95a7-0a0ef5010c28] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1102.303726] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244428, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.334910] env[61907]: INFO nova.compute.resource_tracker [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating resource usage from migration 484d08e2-97b6-4541-a122-65f8f9e81b53 [ 1102.410732] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab32761-8a10-4f45-99f4-733cdc86ddfd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.417982] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f647df57-ec7b-499a-8d08-158b9a3398f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.448982] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1102.451266] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df159c7-14eb-4db1-b1f0-068d02cee9fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.459058] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b404b35-b3e3-406e-8f3b-9018078949fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.472362] env[61907]: DEBUG nova.compute.provider_tree [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1102.584549] env[61907]: DEBUG nova.network.neutron [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Updating instance_info_cache with network_info: [{"id": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "address": "fa:16:3e:c2:6c:03", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce2db7f4-56", "ovs_interfaceid": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.719888] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 53176c9e-d15c-49d5-b4a9-22b780279ecb] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1102.804377] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244428, 'name': CloneVM_Task} progress is 95%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.977816] env[61907]: DEBUG nova.scheduler.client.report [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1103.087654] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.087967] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Instance network_info: |[{"id": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "address": "fa:16:3e:c2:6c:03", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce2db7f4-56", "ovs_interfaceid": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1103.088433] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:6c:03', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce2db7f4-56a8-43ee-904e-1e7c22aa7512', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.096128] env[61907]: DEBUG oslo.service.loopingcall [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.096379] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1103.096621] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ed98de1e-88b9-46e7-a638-9a88108e83de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.117034] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.117034] env[61907]: value = "task-1244429" [ 1103.117034] env[61907]: _type = "Task" [ 1103.117034] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.125356] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244429, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.224227] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5c357582-0b55-4de7-968e-4daa3522dda4] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1103.305122] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244428, 'name': CloneVM_Task, 'duration_secs': 1.26847} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.305367] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created linked-clone VM from snapshot [ 1103.306139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21375a24-3ca6-4bb2-8ba5-cc8a562aa6d3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.313056] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploading image 1e3f6303-b4a1-4cba-a68e-b1e3046f90f0 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1103.335977] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1103.335977] env[61907]: value = "vm-268336" [ 1103.335977] env[61907]: _type = "VirtualMachine" [ 1103.335977] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1103.336292] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a3db62b3-b014-46aa-b664-3e66007409a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.342436] env[61907]: DEBUG nova.compute.manager [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Received event network-changed-ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1103.342667] env[61907]: DEBUG nova.compute.manager [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Refreshing instance network info cache due to event network-changed-ce2db7f4-56a8-43ee-904e-1e7c22aa7512. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1103.342929] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] Acquiring lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.343054] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] Acquired lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.343230] env[61907]: DEBUG nova.network.neutron [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Refreshing network info cache for port ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1103.345778] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease: (returnval){ [ 1103.345778] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c0c9f2-4d1f-c615-486f-5e29bbe45b51" [ 1103.345778] env[61907]: _type = "HttpNfcLease" [ 1103.345778] env[61907]: } obtained for exporting VM: (result){ [ 1103.345778] env[61907]: value = "vm-268336" [ 1103.345778] env[61907]: _type = "VirtualMachine" [ 1103.345778] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1103.346028] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the lease: (returnval){ [ 1103.346028] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c0c9f2-4d1f-c615-486f-5e29bbe45b51" [ 1103.346028] env[61907]: _type = "HttpNfcLease" [ 1103.346028] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1103.353926] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1103.353926] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c0c9f2-4d1f-c615-486f-5e29bbe45b51" [ 1103.353926] env[61907]: _type = "HttpNfcLease" [ 1103.353926] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1103.482964] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.159s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.483205] env[61907]: INFO nova.compute.manager [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Migrating [ 1103.627517] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244429, 'name': CreateVM_Task, 'duration_secs': 0.318859} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.627700] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1103.628422] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.628611] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.628947] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1103.629215] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0837bed-a80a-4e33-9658-102f7c55ba7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.633519] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1103.633519] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f556fa-e25d-4d38-5e90-9f4e48635ebd" [ 1103.633519] env[61907]: _type = "Task" [ 1103.633519] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.641128] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f556fa-e25d-4d38-5e90-9f4e48635ebd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.727973] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 37ff6c54-6b79-4a9c-bc16-29d974185d3e] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1103.856455] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1103.856455] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c0c9f2-4d1f-c615-486f-5e29bbe45b51" [ 1103.856455] env[61907]: _type = "HttpNfcLease" [ 1103.856455] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1103.856729] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1103.856729] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52c0c9f2-4d1f-c615-486f-5e29bbe45b51" [ 1103.856729] env[61907]: _type = "HttpNfcLease" [ 1103.856729] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1103.857441] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda969e2-b027-4f8c-941a-49c58421f636 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.864386] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1103.864559] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1103.950835] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-8f71bf60-fb8e-40cc-b55e-7434cb23cbee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.997600] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.997797] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.997976] env[61907]: DEBUG nova.network.neutron [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1104.083832] env[61907]: DEBUG nova.network.neutron [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Updated VIF entry in instance network info cache for port ce2db7f4-56a8-43ee-904e-1e7c22aa7512. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1104.084220] env[61907]: DEBUG nova.network.neutron [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Updating instance_info_cache with network_info: [{"id": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "address": "fa:16:3e:c2:6c:03", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce2db7f4-56", "ovs_interfaceid": "ce2db7f4-56a8-43ee-904e-1e7c22aa7512", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.144350] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f556fa-e25d-4d38-5e90-9f4e48635ebd, 'name': SearchDatastore_Task, 'duration_secs': 0.010145} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.144705] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.144954] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.145212] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.145394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.145628] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.145876] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f43064ce-6e15-414c-aa38-8d5ede1717a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.154074] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.154292] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.155069] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d70a2b7-288e-41a7-8b78-c09c5bb34ae2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.160308] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1104.160308] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ce4b08-e743-b32e-42d8-49e20a5491e1" [ 1104.160308] env[61907]: _type = "Task" [ 1104.160308] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.167949] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ce4b08-e743-b32e-42d8-49e20a5491e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.231265] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: e52648ac-e068-4a63-baa0-a1c34df52197] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1104.586384] env[61907]: DEBUG oslo_concurrency.lockutils [req-ef34766b-eb4e-410c-bd38-15eb53d6e737 req-77c389aa-9cb7-498e-8164-9fb210c63366 service nova] Releasing lock "refresh_cache-0e9c2580-cf31-45fd-b5a7-ee13291056aa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.671170] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ce4b08-e743-b32e-42d8-49e20a5491e1, 'name': SearchDatastore_Task, 'duration_secs': 0.009117} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.672519] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da64d630-68ae-4f40-b5d4-cebe9060fe0e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.677187] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1104.677187] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b23fab-89d9-24ee-2210-63ec6af8a9ba" [ 1104.677187] env[61907]: _type = "Task" [ 1104.677187] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.686613] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b23fab-89d9-24ee-2210-63ec6af8a9ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.731692] env[61907]: DEBUG nova.network.neutron [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.734825] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: d32315e1-68b9-49ee-9393-8b2f9cbaf645] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1105.188349] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b23fab-89d9-24ee-2210-63ec6af8a9ba, 'name': SearchDatastore_Task, 'duration_secs': 0.019863} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.188977] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.189347] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 0e9c2580-cf31-45fd-b5a7-ee13291056aa/0e9c2580-cf31-45fd-b5a7-ee13291056aa.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1105.189619] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d301ab3-db38-4df0-8ab2-57e293cb4e97 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.196264] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1105.196264] env[61907]: value = "task-1244431" [ 1105.196264] env[61907]: _type = "Task" [ 1105.196264] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.206182] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244431, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.234076] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.238486] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: aec01156-431b-447e-88ef-cb8d254c6341] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1105.710384] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244431, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.746051] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 0bfdb2d0-8388-4be8-a2ee-743c029db6c3] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1106.208146] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244431, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567661} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.208463] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 0e9c2580-cf31-45fd-b5a7-ee13291056aa/0e9c2580-cf31-45fd-b5a7-ee13291056aa.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1106.208715] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1106.209026] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1fc32eef-e8db-4645-9d7b-a32f06590b1a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.215307] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1106.215307] env[61907]: value = "task-1244432" [ 1106.215307] env[61907]: _type = "Task" [ 1106.215307] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.224125] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244432, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.251533] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2c1212a7-df13-49b7-b3b3-24fb59b6789d] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1106.726521] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244432, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073788} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.726810] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1106.727641] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4bcb52-b51a-447f-adef-330b8add04b7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.750125] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] 0e9c2580-cf31-45fd-b5a7-ee13291056aa/0e9c2580-cf31-45fd-b5a7-ee13291056aa.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.750485] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f0c6c229-6ccb-4065-838f-04036e9a03e6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.768588] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f028beaf-4185-40d7-a730-abe91209f9d8] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1106.771604] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb30091-b83f-4201-a093-4c3daf944e3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.776036] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1106.776036] env[61907]: value = "task-1244433" [ 1106.776036] env[61907]: _type = "Task" [ 1106.776036] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.792517] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1106.804429] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244433, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.275084] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 41c680fb-5450-43f3-9acb-5218e57a7da9] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1107.287049] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244433, 'name': ReconfigVM_Task, 'duration_secs': 0.335926} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.287049] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Reconfigured VM instance instance-00000069 to attach disk [datastore2] 0e9c2580-cf31-45fd-b5a7-ee13291056aa/0e9c2580-cf31-45fd-b5a7-ee13291056aa.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.287455] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-293b1306-6ffc-4dd3-bdf0-a063827f3029 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.293250] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1107.293250] env[61907]: value = "task-1244434" [ 1107.293250] env[61907]: _type = "Task" [ 1107.293250] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.299013] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1107.302740] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-483114f8-a9fb-4272-953d-c3a7fabd0f91 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.304242] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244434, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.311644] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1107.311644] env[61907]: value = "task-1244435" [ 1107.311644] env[61907]: _type = "Task" [ 1107.311644] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.320818] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.782358] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ddcbf498-fd27-41f2-bbcc-23a8f9827823] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1107.803064] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244434, 'name': Rename_Task, 'duration_secs': 0.18297} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.803321] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1107.803586] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5aec992f-d225-4970-b5c1-7bc6318cf9f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.811489] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1107.811489] env[61907]: value = "task-1244436" [ 1107.811489] env[61907]: _type = "Task" [ 1107.811489] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.822796] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244436, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.825872] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244435, 'name': PowerOffVM_Task, 'duration_secs': 0.186965} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.826134] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1107.826358] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1108.286211] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 4c70ef23-fd26-4e90-b227-13aa050ff46d] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1108.322538] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244436, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.332770] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1108.333035] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1108.333221] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1108.333520] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1108.333636] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1108.333756] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1108.333970] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1108.334171] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1108.334406] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1108.334594] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1108.334779] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1108.340237] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c40b7db-59f1-423e-936a-2efc38874b7b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.356692] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1108.356692] env[61907]: value = "task-1244437" [ 1108.356692] env[61907]: _type = "Task" [ 1108.356692] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.364641] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244437, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.790255] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: da490943-511b-4776-8f16-4f51c3b055c4] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1108.822583] env[61907]: DEBUG oslo_vmware.api [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244436, 'name': PowerOnVM_Task, 'duration_secs': 0.545358} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.822821] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.823034] env[61907]: INFO nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Took 7.62 seconds to spawn the instance on the hypervisor. [ 1108.823225] env[61907]: DEBUG nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1108.823998] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36f9f30b-21ee-4411-9145-0bd487817d56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.867164] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244437, 'name': ReconfigVM_Task, 'duration_secs': 0.326738} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.867500] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1109.295383] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 23526967-cd8f-4581-b9c5-1c270d385163] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1109.341034] env[61907]: INFO nova.compute.manager [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Took 12.37 seconds to build instance. [ 1109.373912] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1109.374212] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1109.374395] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1109.374592] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1109.374747] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1109.374901] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1109.375133] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1109.375304] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1109.375498] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1109.375664] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1109.375846] env[61907]: DEBUG nova.virt.hardware [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1109.381198] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1109.381746] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae332c4c-af0a-48f7-b1a3-9409314caf8e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.400795] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1109.400795] env[61907]: value = "task-1244438" [ 1109.400795] env[61907]: _type = "Task" [ 1109.400795] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.408770] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244438, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.798790] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 2ab93ab9-bf69-4525-8df6-eef83dd24bc1] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1109.842833] env[61907]: DEBUG oslo_concurrency.lockutils [None req-49e63eab-5f53-426c-a54a-d7c39c17af5c tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.883s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.912898] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244438, 'name': ReconfigVM_Task, 'duration_secs': 0.163046} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.913362] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1109.914292] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b564f3b-4e9c-4cc9-aac4-92834fdc5ee6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.936407] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1109.936724] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95b783e1-572d-491f-8004-e47007503fc5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.954416] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1109.954416] env[61907]: value = "task-1244439" [ 1109.954416] env[61907]: _type = "Task" [ 1109.954416] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.962259] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244439, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.302340] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 92f27804-8974-40c4-9663-b2b72f0bb8e0] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1110.464672] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244439, 'name': ReconfigVM_Task, 'duration_secs': 0.273421} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.464898] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1110.465198] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1110.804513] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "bc6d118d-41c7-4a7a-9b75-14919844899a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.804767] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.805974] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: e1f8f88b-7179-4e92-83cd-a1027ff9b0a2] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1110.972110] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e32c07-190a-4d5a-8cf6-6f72cd9bd4e5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.990888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ee7272-1e0a-4dd7-8ffc-3df64823c26d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.008486] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1111.308286] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1111.311719] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5c908f9c-7efb-4588-b6bf-70babb173767] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1111.550435] env[61907]: DEBUG nova.network.neutron [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1111.815359] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: e5d4890e-be1b-4b28-9721-89ef678c1182] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1111.832249] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.832522] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.834108] env[61907]: INFO nova.compute.claims [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1112.320649] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 7ba2aa4c-3bd2-4451-a3f3-fcdfba0ed111] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1112.573647] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.573846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.574054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.824554] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f157c0db-054b-4d99-803c-11aa40cf0670] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1112.940947] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-312f8a35-228b-4d66-abad-4487b6182177 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.949754] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b54435-46c5-481b-a262-9cb41f6e8094 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.981015] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad013587-bf9b-4e22-9f5f-6b9223a01eda {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.989210] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce024304-d09d-41a2-9b0b-26b0bc1c979c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.003023] env[61907]: DEBUG nova.compute.provider_tree [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.196825] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1113.197859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85597f6f-5d65-49d0-afef-9dc615575875 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.203826] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1113.204007] env[61907]: ERROR oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk due to incomplete transfer. [ 1113.204255] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bb68e389-bb0e-4421-9d43-d9c77cf285bf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.211596] env[61907]: DEBUG oslo_vmware.rw_handles [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52d53a23-7ee4-b9b3-a553-efa7d4e605b2/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1113.211799] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploaded image 1e3f6303-b4a1-4cba-a68e-b1e3046f90f0 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1113.214155] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1113.214393] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d88a0814-2a11-41cb-9b8a-3c5df001fd2e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.220260] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1113.220260] env[61907]: value = "task-1244440" [ 1113.220260] env[61907]: _type = "Task" [ 1113.220260] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.228351] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244440, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.328229] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 52404147-0375-4f9c-9e1e-4a52001df7cb] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1113.507163] env[61907]: DEBUG nova.scheduler.client.report [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1113.620082] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1113.620300] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1113.620485] env[61907]: DEBUG nova.network.neutron [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1113.729715] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244440, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.831854] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ce864af6-aef6-4044-be64-8440ba175438] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1114.013027] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.013170] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1114.231401] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244440, 'name': Destroy_Task} progress is 33%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.336560] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 08647e48-9e24-4d7c-a9fd-4066317a7250] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1114.385707] env[61907]: DEBUG nova.network.neutron [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1114.518864] env[61907]: DEBUG nova.compute.utils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1114.520767] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1114.521065] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1114.561540] env[61907]: DEBUG nova.policy [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1114.731785] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244440, 'name': Destroy_Task, 'duration_secs': 1.324791} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.732062] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroyed the VM [ 1114.732300] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1114.732550] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bfcea673-db15-42e9-9a9e-638ccb00ab76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.738688] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1114.738688] env[61907]: value = "task-1244441" [ 1114.738688] env[61907]: _type = "Task" [ 1114.738688] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.745853] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244441, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.803584] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Successfully created port: ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1114.839439] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5b20f363-dfad-4c2b-b757-4da4e5bd0b99] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1114.888127] env[61907]: DEBUG oslo_concurrency.lockutils [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.028378] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1115.248536] env[61907]: DEBUG oslo_vmware.api [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244441, 'name': RemoveSnapshot_Task, 'duration_secs': 0.47983} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.248807] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1115.249061] env[61907]: INFO nova.compute.manager [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 15.06 seconds to snapshot the instance on the hypervisor. [ 1115.342715] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 1ef8594c-5d9a-491f-89f6-e5d7f35ce751] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1115.414597] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bae42e-8a4f-4c99-830b-40b1118f316c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.434043] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a99c524d-648c-445b-a7e6-4daca7f5c0a7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.441347] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1115.791396] env[61907]: DEBUG nova.compute.manager [None req-954d78c7-3bc3-4e5c-bac0-cbd8777756ea tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Found 2 images (rotation: 2) {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 1115.846696] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 865d386d-c5f0-439b-b50b-b613993ec9b8] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1115.947296] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1115.947633] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-440bc3e8-d94b-489c-9bd7-36c4351da80b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.954384] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1115.954384] env[61907]: value = "task-1244442" [ 1115.954384] env[61907]: _type = "Task" [ 1115.954384] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.962011] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.036513] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1116.060522] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1116.060853] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1116.061066] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1116.061356] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1116.061558] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1116.061740] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1116.062015] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1116.062319] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1116.062412] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1116.062587] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1116.062773] env[61907]: DEBUG nova.virt.hardware [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1116.063734] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f876fe1-2476-4f95-8cb5-39247859a993 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.072045] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52042894-e3c7-4388-8246-4888ed774b28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.193923] env[61907]: DEBUG nova.compute.manager [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Received event network-vif-plugged-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1116.194205] env[61907]: DEBUG oslo_concurrency.lockutils [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] Acquiring lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.194489] env[61907]: DEBUG oslo_concurrency.lockutils [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.194639] env[61907]: DEBUG oslo_concurrency.lockutils [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1116.194823] env[61907]: DEBUG nova.compute.manager [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] No waiting events found dispatching network-vif-plugged-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1116.194994] env[61907]: WARNING nova.compute.manager [req-a2070a14-6ee9-49d9-864a-0ae8ba8f662e req-a175ef72-1940-4ae4-aae3-663642363d61 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Received unexpected event network-vif-plugged-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 for instance with vm_state building and task_state spawning. [ 1116.291973] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Successfully updated port: ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1116.350144] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: a6a117b3-7cc4-43c5-8202-12514f695e48] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1116.464253] env[61907]: DEBUG oslo_vmware.api [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244442, 'name': PowerOnVM_Task, 'duration_secs': 0.372685} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.464561] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1116.464757] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-df277843-3f58-49fd-bf69-11a15dd35830 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance '378b5c99-5dbc-4657-8504-58d20dc6c9ea' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1116.795138] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.795307] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.795495] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1116.853120] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 9209400d-d5b8-4b96-9022-f8a4c7a30471] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1117.218563] env[61907]: DEBUG nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1117.219539] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b53274a-d6b8-478c-843d-c801ed286786 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.326908] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1117.356738] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 691fff4a-9dbd-47f2-8908-69057a2bb6fe] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1117.469716] env[61907]: DEBUG nova.network.neutron [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Updating instance_info_cache with network_info: [{"id": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "address": "fa:16:3e:c3:d6:43", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf7b56a-1f", "ovs_interfaceid": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.730067] env[61907]: INFO nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] instance snapshotting [ 1117.730716] env[61907]: DEBUG nova.objects.instance [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.860255] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: da7ddc92-f3e3-4b4d-a14b-13fd25a72314] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1117.972627] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.972938] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Instance network_info: |[{"id": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "address": "fa:16:3e:c3:d6:43", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf7b56a-1f", "ovs_interfaceid": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1117.973397] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:d6:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1117.980903] env[61907]: DEBUG oslo.service.loopingcall [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1117.990100] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1117.990559] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57a1fb6a-253e-42a0-889d-f2e217e9863c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.013696] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1118.013696] env[61907]: value = "task-1244443" [ 1118.013696] env[61907]: _type = "Task" [ 1118.013696] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.024284] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244443, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.227051] env[61907]: DEBUG nova.compute.manager [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Received event network-changed-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1118.227338] env[61907]: DEBUG nova.compute.manager [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Refreshing instance network info cache due to event network-changed-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1118.227564] env[61907]: DEBUG oslo_concurrency.lockutils [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] Acquiring lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.227827] env[61907]: DEBUG oslo_concurrency.lockutils [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] Acquired lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.228241] env[61907]: DEBUG nova.network.neutron [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Refreshing network info cache for port ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1118.235719] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-941df7cc-0a2c-4344-bbdf-88d5df6c3ac9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.256860] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24b956fa-62be-43c0-8be5-27505ca4de24 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.363434] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1118.389824] env[61907]: DEBUG nova.network.neutron [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Port 9fa8da7e-09f8-42c1-8705-6a29e9024ad5 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1118.390111] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.390274] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.390445] env[61907]: DEBUG nova.network.neutron [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1118.523996] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244443, 'name': CreateVM_Task, 'duration_secs': 0.299965} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1118.524204] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1118.525017] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.525247] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.525635] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1118.525928] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a778a064-8c5a-4a83-bc5c-328f914d242c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.530703] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1118.530703] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f69a7e-4718-bb5d-3e4d-300d9a443031" [ 1118.530703] env[61907]: _type = "Task" [ 1118.530703] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.539229] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f69a7e-4718-bb5d-3e4d-300d9a443031, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.766816] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1118.767149] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ab74b374-176f-4eae-8f63-63391403512e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.774649] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1118.774649] env[61907]: value = "task-1244444" [ 1118.774649] env[61907]: _type = "Task" [ 1118.774649] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.782432] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244444, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.934802] env[61907]: DEBUG nova.network.neutron [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Updated VIF entry in instance network info cache for port ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1118.934802] env[61907]: DEBUG nova.network.neutron [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Updating instance_info_cache with network_info: [{"id": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "address": "fa:16:3e:c3:d6:43", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapccf7b56a-1f", "ovs_interfaceid": "ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.048401] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f69a7e-4718-bb5d-3e4d-300d9a443031, 'name': SearchDatastore_Task, 'duration_secs': 0.010054} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.048829] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.049200] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1119.049502] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1119.049707] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1119.049942] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1119.050270] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-620734aa-9506-4212-a4f7-460ce444d6fa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.059057] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1119.059057] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1119.062032] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84a5485e-5639-4ecf-b4ae-b6a48ee635fd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.066664] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1119.066664] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5288aca1-173e-00c8-eadf-65e8b39cabf0" [ 1119.066664] env[61907]: _type = "Task" [ 1119.066664] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.074187] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5288aca1-173e-00c8-eadf-65e8b39cabf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.157888] env[61907]: DEBUG nova.network.neutron [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.284949] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244444, 'name': CreateSnapshot_Task, 'duration_secs': 0.473751} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.285197] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1119.285906] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e6563f-e86e-4b17-af46-4867fb50415b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.439179] env[61907]: DEBUG oslo_concurrency.lockutils [req-c256168c-e98a-4f3d-a41a-210b5004d229 req-9762846d-8267-48a1-91f7-0500568cadf9 service nova] Releasing lock "refresh_cache-bc6d118d-41c7-4a7a-9b75-14919844899a" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.578061] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5288aca1-173e-00c8-eadf-65e8b39cabf0, 'name': SearchDatastore_Task, 'duration_secs': 0.008782} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.578804] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16f65282-065a-46f6-a47d-ef3f21d866ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.585043] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1119.585043] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52468ca4-03c7-1a4f-38bc-e76c01af29ff" [ 1119.585043] env[61907]: _type = "Task" [ 1119.585043] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.591916] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52468ca4-03c7-1a4f-38bc-e76c01af29ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.660783] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1119.803668] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1119.804320] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fd3e4086-d653-422d-9f4b-4a542ce1aa5d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.813125] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1119.813125] env[61907]: value = "task-1244445" [ 1119.813125] env[61907]: _type = "Task" [ 1119.813125] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.820396] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244445, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.096851] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52468ca4-03c7-1a4f-38bc-e76c01af29ff, 'name': SearchDatastore_Task, 'duration_secs': 0.016727} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1120.097218] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.097581] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] bc6d118d-41c7-4a7a-9b75-14919844899a/bc6d118d-41c7-4a7a-9b75-14919844899a.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1120.097912] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6f1ac0a4-5a53-40ba-9c43-f68b808fafbb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.106122] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1120.106122] env[61907]: value = "task-1244446" [ 1120.106122] env[61907]: _type = "Task" [ 1120.106122] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1120.116472] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244446, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.165038] env[61907]: DEBUG nova.compute.manager [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61907) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1120.165258] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1120.165546] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.323127] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244445, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.616010] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244446, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.668825] env[61907]: DEBUG nova.objects.instance [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'migration_context' on Instance uuid 378b5c99-5dbc-4657-8504-58d20dc6c9ea {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.823529] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244445, 'name': CloneVM_Task} progress is 95%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.118975] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244446, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546043} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.119369] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] bc6d118d-41c7-4a7a-9b75-14919844899a/bc6d118d-41c7-4a7a-9b75-14919844899a.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1121.119593] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1121.119849] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c9c830de-754f-4f30-be79-e534bce588d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.125260] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1121.125260] env[61907]: value = "task-1244447" [ 1121.125260] env[61907]: _type = "Task" [ 1121.125260] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.132230] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244447, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.268456] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535f84a4-7174-4ea0-b484-f9ef4415da75 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.276041] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d905d139-bc0f-4bcb-a9a6-0ecefac8e2d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.305724] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8c4be7-7f86-494b-917b-9b8bfd537fc5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.312966] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3041e0e1-e213-450f-b2a8-cbef27cd593e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.323833] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244445, 'name': CloneVM_Task, 'duration_secs': 1.300873} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.330817] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Created linked-clone VM from snapshot [ 1121.331276] env[61907]: DEBUG nova.compute.provider_tree [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1121.332785] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-544c4d4e-9796-4ba2-927b-6f3f173dfbe3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.339510] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploading image 45ae0561-6dc2-442d-a1c3-4d09e4b6e8e2 {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1121.357019] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.357284] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.366384] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1121.366384] env[61907]: value = "vm-268340" [ 1121.366384] env[61907]: _type = "VirtualMachine" [ 1121.366384] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1121.366752] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8a805844-5a81-4359-8367-530c46c61fdb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.373975] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease: (returnval){ [ 1121.373975] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526984f1-aad3-c673-9be4-dc122537cdc2" [ 1121.373975] env[61907]: _type = "HttpNfcLease" [ 1121.373975] env[61907]: } obtained for exporting VM: (result){ [ 1121.373975] env[61907]: value = "vm-268340" [ 1121.373975] env[61907]: _type = "VirtualMachine" [ 1121.373975] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1121.374319] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the lease: (returnval){ [ 1121.374319] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526984f1-aad3-c673-9be4-dc122537cdc2" [ 1121.374319] env[61907]: _type = "HttpNfcLease" [ 1121.374319] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1121.381177] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1121.381177] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526984f1-aad3-c673-9be4-dc122537cdc2" [ 1121.381177] env[61907]: _type = "HttpNfcLease" [ 1121.381177] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1121.634452] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244447, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065106} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.634768] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1121.635426] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5905c944-cca4-4399-80a1-2dd6e75f8acd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.656683] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] bc6d118d-41c7-4a7a-9b75-14919844899a/bc6d118d-41c7-4a7a-9b75-14919844899a.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1121.656920] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5cc76b1-be59-4fd5-9dfa-930e7916de84 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.675734] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1121.675734] env[61907]: value = "task-1244449" [ 1121.675734] env[61907]: _type = "Task" [ 1121.675734] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.684393] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244449, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.837107] env[61907]: DEBUG nova.scheduler.client.report [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1121.862202] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1121.862374] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1121.862374] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1121.884007] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1121.884007] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526984f1-aad3-c673-9be4-dc122537cdc2" [ 1121.884007] env[61907]: _type = "HttpNfcLease" [ 1121.884007] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1121.884569] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1121.884569] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526984f1-aad3-c673-9be4-dc122537cdc2" [ 1121.884569] env[61907]: _type = "HttpNfcLease" [ 1121.884569] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1121.885274] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-588c3fce-20ae-48ac-ae07-d89dbf89a6c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.892460] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1121.892640] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1121.981466] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-03314844-b868-4435-b377-346f61369480 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.185612] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244449, 'name': ReconfigVM_Task, 'duration_secs': 0.478782} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.185961] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Reconfigured VM instance instance-0000006a to attach disk [datastore2] bc6d118d-41c7-4a7a-9b75-14919844899a/bc6d118d-41c7-4a7a-9b75-14919844899a.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1122.186702] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72c21a26-4ad7-45c6-af3d-9f73fd1c1a3e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.193171] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1122.193171] env[61907]: value = "task-1244450" [ 1122.193171] env[61907]: _type = "Task" [ 1122.193171] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.201771] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244450, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.366723] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Skipping network cache update for instance because it is Building. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10284}} [ 1122.397024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.397024] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.397024] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1122.397024] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1122.703847] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244450, 'name': Rename_Task, 'duration_secs': 0.126781} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.704321] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1122.704677] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d8a3ebfc-d159-418d-b8ef-ced0f7a0fe3d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.711187] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1122.711187] env[61907]: value = "task-1244451" [ 1122.711187] env[61907]: _type = "Task" [ 1122.711187] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.720575] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.847645] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.682s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.224629] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244451, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.722180] env[61907]: DEBUG oslo_vmware.api [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244451, 'name': PowerOnVM_Task, 'duration_secs': 0.560881} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.722559] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1123.722802] env[61907]: INFO nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Took 7.69 seconds to spawn the instance on the hypervisor. [ 1123.722996] env[61907]: DEBUG nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1123.723817] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2690de1-7d24-4001-a9ca-3b0bae6a33c3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.141040] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [{"id": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "address": "fa:16:3e:f2:c0:21", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.213", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "taped422e68-2f", "ovs_interfaceid": "ed422e68-2fc3-4015-8d2a-716f55d3ea41", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.241101] env[61907]: INFO nova.compute.manager [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Took 12.43 seconds to build instance. [ 1124.389465] env[61907]: INFO nova.compute.manager [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Swapping old allocation on dict_keys(['91bca385-a423-4ca4-9da0-aeb4615e22d3']) held by migration 484d08e2-97b6-4541-a122-65f8f9e81b53 for instance [ 1124.416033] env[61907]: DEBUG nova.scheduler.client.report [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Overwriting current allocation {'allocations': {'91bca385-a423-4ca4-9da0-aeb4615e22d3': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 139}}, 'project_id': '7652e98cde994af28b7bac0b81547474', 'user_id': '88b29a917e03445281f16fba6d536280', 'consumer_generation': 1} on consumer 378b5c99-5dbc-4657-8504-58d20dc6c9ea {{(pid=61907) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1124.497543] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.497788] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.497975] env[61907]: DEBUG nova.network.neutron [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1124.643640] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.643867] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1124.644101] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.644266] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.644435] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.644605] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.644749] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.644895] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.645035] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1124.645185] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.743145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-9c5ca1f3-82f4-43c0-a041-667f15882996 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.938s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.149697] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.149928] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.150109] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.150265] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1125.151213] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82e1cb7-53e6-49d6-b5e7-b9bbe9ca70ab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.160075] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3e2be8-109c-40ff-bd41-954c513715d6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.174071] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd97020b-cc50-4dc2-b014-355ede3ce293 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.180861] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e18e3e4-5345-4d8a-bab6-aec2492448b0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.212293] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180707MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1125.212477] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.212657] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.246060] env[61907]: DEBUG nova.network.neutron [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [{"id": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "address": "fa:16:3e:7d:de:7f", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9fa8da7e-09", "ovs_interfaceid": "9fa8da7e-09f8-42c1-8705-6a29e9024ad5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.448286] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "bc6d118d-41c7-4a7a-9b75-14919844899a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.448652] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.448906] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.449130] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.449335] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.451618] env[61907]: INFO nova.compute.manager [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Terminating instance [ 1125.749648] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-378b5c99-5dbc-4657-8504-58d20dc6c9ea" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.750037] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.750328] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c94a47e5-a367-464b-9725-ed081083509a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.758318] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1125.758318] env[61907]: value = "task-1244452" [ 1125.758318] env[61907]: _type = "Task" [ 1125.758318] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.769014] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244452, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.955982] env[61907]: DEBUG nova.compute.manager [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1125.955982] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1125.957078] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3b53ab-ee81-4a6b-bab1-c37d6600b448 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.971766] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1125.972161] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6df1c079-2a83-4174-8dca-8c36656f7c71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.978838] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1125.978838] env[61907]: value = "task-1244453" [ 1125.978838] env[61907]: _type = "Task" [ 1125.978838] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.989670] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244453, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.242039] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242205] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f7959e23-8879-4285-93da-804dd3115f65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242333] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance b2c5689e-f930-4548-b061-fd2974d44575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242482] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 0e9c2580-cf31-45fd-b5a7-ee13291056aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242612] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 378b5c99-5dbc-4657-8504-58d20dc6c9ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242727] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance bc6d118d-41c7-4a7a-9b75-14919844899a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1126.242937] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1126.243106] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1126.269878] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244452, 'name': PowerOffVM_Task, 'duration_secs': 0.461263} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.272602] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.273318] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.273563] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.273746] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.273936] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.274116] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.274294] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.274528] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.274726] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.274906] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.275089] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.275290] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.280661] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9df4789-3622-4dc0-b055-3e7ac9b60994 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.299382] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1126.299382] env[61907]: value = "task-1244454" [ 1126.299382] env[61907]: _type = "Task" [ 1126.299382] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.308810] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244454, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.337050] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f8e6d1-ef43-4e11-b621-9488903e420a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.344456] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2ff7c5-de10-49fd-abc3-77b57ba060bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.375807] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f67b6f-2186-4712-97fa-d4b69d3af040 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.383787] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b8efe1-90da-4bbe-a425-426b6542c02a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.397237] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.488798] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244453, 'name': PowerOffVM_Task, 'duration_secs': 0.24698} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.489104] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.489297] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1126.489573] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25bb22a9-4623-4d71-bd7d-af5a20d28141 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.549205] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1126.549514] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1126.549711] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] bc6d118d-41c7-4a7a-9b75-14919844899a {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1126.549994] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-978bedd1-db20-4c10-bc1d-b5082de90094 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.556196] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1126.556196] env[61907]: value = "task-1244456" [ 1126.556196] env[61907]: _type = "Task" [ 1126.556196] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.564807] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244456, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.808945] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244454, 'name': ReconfigVM_Task, 'duration_secs': 0.130093} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.809813] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc4908b-e965-4d0c-bc26-94e37af91b72 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.828198] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1126.828494] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1126.828682] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1126.828880] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1126.829067] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1126.829224] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1126.829451] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1126.829626] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1126.829820] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1126.829984] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1126.830204] env[61907]: DEBUG nova.virt.hardware [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1126.831098] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9fe1695b-9b16-4b43-8e4e-ef0e098d6966 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.836543] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1126.836543] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f5b7f1-6ff1-34b9-cb1e-434b1bfc1d03" [ 1126.836543] env[61907]: _type = "Task" [ 1126.836543] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.844110] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f5b7f1-6ff1-34b9-cb1e-434b1bfc1d03, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.900619] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1127.066516] env[61907]: DEBUG oslo_vmware.api [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244456, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.180967} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.066780] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.066971] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1127.067179] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1127.067362] env[61907]: INFO nova.compute.manager [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1127.067620] env[61907]: DEBUG oslo.service.loopingcall [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.067819] env[61907]: DEBUG nova.compute.manager [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1127.067917] env[61907]: DEBUG nova.network.neutron [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1127.314071] env[61907]: DEBUG nova.compute.manager [req-2ca47a6e-c55f-46b0-be93-28f90ec837da req-c6417de7-16e4-4c0a-8148-b5fbca8c0a77 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Received event network-vif-deleted-ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1127.314302] env[61907]: INFO nova.compute.manager [req-2ca47a6e-c55f-46b0-be93-28f90ec837da req-c6417de7-16e4-4c0a-8148-b5fbca8c0a77 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Neutron deleted interface ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24; detaching it from the instance and deleting it from the info cache [ 1127.314506] env[61907]: DEBUG nova.network.neutron [req-2ca47a6e-c55f-46b0-be93-28f90ec837da req-c6417de7-16e4-4c0a-8148-b5fbca8c0a77 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.347333] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f5b7f1-6ff1-34b9-cb1e-434b1bfc1d03, 'name': SearchDatastore_Task, 'duration_secs': 0.041019} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.352821] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1127.353407] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97a7da2b-5d08-43b0-b371-62e8cc420813 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.372286] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1127.372286] env[61907]: value = "task-1244457" [ 1127.372286] env[61907]: _type = "Task" [ 1127.372286] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.380734] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244457, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.405481] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1127.405693] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.193s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.789794] env[61907]: DEBUG nova.network.neutron [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1127.816736] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e58485ec-5741-4342-80bf-116baacd4214 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.827049] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6a2de3-fb7a-4a3f-ae0c-9c49c5978651 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.854163] env[61907]: DEBUG nova.compute.manager [req-2ca47a6e-c55f-46b0-be93-28f90ec837da req-c6417de7-16e4-4c0a-8148-b5fbca8c0a77 service nova] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Detach interface failed, port_id=ccf7b56a-1ff0-4779-b2bd-f7d37fc1ab24, reason: Instance bc6d118d-41c7-4a7a-9b75-14919844899a could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1127.881254] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244457, 'name': ReconfigVM_Task, 'duration_secs': 0.196878} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.881559] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1127.882354] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b743d2-c1c2-4845-bacd-9f413d8eeeae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.904407] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.904730] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a60354a6-d308-4dc5-9d0a-c9297bdd01ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.922341] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1127.922341] env[61907]: value = "task-1244458" [ 1127.922341] env[61907]: _type = "Task" [ 1127.922341] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.930743] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.292731] env[61907]: INFO nova.compute.manager [-] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Took 1.22 seconds to deallocate network for instance. [ 1128.432619] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244458, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.800846] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.801221] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.801492] env[61907]: DEBUG nova.objects.instance [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid bc6d118d-41c7-4a7a-9b75-14919844899a {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.932994] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244458, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.399013] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be3dba2-91fe-4065-924d-87e187de5e64 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.406612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91fd673-8202-481e-80f8-691dcb558310 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.441852] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5c706c-26cd-49fd-b4c3-bd59d64cb1d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.449387] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244458, 'name': ReconfigVM_Task, 'duration_secs': 1.423534} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.451525] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea/378b5c99-5dbc-4657-8504-58d20dc6c9ea.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.452431] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f86264f-809d-4224-aed0-a9c5393ca115 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.455797] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b878d052-0503-4b40-8874-0ae8b426f575 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.483890] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8505a9-f289-4f64-8eb7-56634a99948a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.486860] env[61907]: DEBUG nova.compute.provider_tree [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.504214] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b31cead6-4ae5-4691-9a89-52443e2a229b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.523888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978f6612-b2f5-4a14-8b96-782cbace6871 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.530697] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1129.530950] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d08091bd-0271-4258-bab7-9b0c81b3edf1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.536713] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1129.536713] env[61907]: value = "task-1244459" [ 1129.536713] env[61907]: _type = "Task" [ 1129.536713] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.543734] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244459, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.570982] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1129.571920] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6eadae-54fc-4455-86d2-5a69c2327b73 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.577719] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1129.577865] env[61907]: ERROR oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk due to incomplete transfer. [ 1129.578122] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-25f2564f-b064-4133-acea-ae14a98a03ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.584621] env[61907]: DEBUG oslo_vmware.rw_handles [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/525879c6-3f7e-0eef-0158-816534ec779a/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1129.584814] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Uploaded image 45ae0561-6dc2-442d-a1c3-4d09e4b6e8e2 to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1129.587176] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1129.587432] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-36bf908e-5485-499a-9600-44889ae9af77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.594609] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1129.594609] env[61907]: value = "task-1244460" [ 1129.594609] env[61907]: _type = "Task" [ 1129.594609] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.601717] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244460, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.991029] env[61907]: DEBUG nova.scheduler.client.report [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1130.046054] env[61907]: DEBUG oslo_vmware.api [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244459, 'name': PowerOnVM_Task, 'duration_secs': 0.382485} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.046339] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1130.103387] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244460, 'name': Destroy_Task, 'duration_secs': 0.345653} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.103659] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroyed the VM [ 1130.103898] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1130.104176] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6dde4392-1d82-4fdc-a509-46816d168f56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.110910] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1130.110910] env[61907]: value = "task-1244461" [ 1130.110910] env[61907]: _type = "Task" [ 1130.110910] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.118147] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244461, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.495663] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.516326] env[61907]: INFO nova.scheduler.client.report [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance bc6d118d-41c7-4a7a-9b75-14919844899a [ 1130.620813] env[61907]: DEBUG oslo_vmware.api [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244461, 'name': RemoveSnapshot_Task, 'duration_secs': 0.47852} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.621110] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1130.621347] env[61907]: INFO nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 12.39 seconds to snapshot the instance on the hypervisor. [ 1131.024026] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5e64607e-8374-4f6a-98e1-fdb6c98b643b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "bc6d118d-41c7-4a7a-9b75-14919844899a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.575s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.056285] env[61907]: INFO nova.compute.manager [None req-7337de6c-ea66-4b57-becb-7b514c07385f tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance to original state: 'active' [ 1131.161685] env[61907]: DEBUG nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Found 3 images (rotation: 2) {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4897}} [ 1131.161914] env[61907]: DEBUG nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Rotating out 1 backups {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4905}} [ 1131.162098] env[61907]: DEBUG nova.compute.manager [None req-9491a091-ac7a-4dad-bb0c-cb54508345ab tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleting image 4f4d5745-4c73-47d1-8075-14da12669c40 {{(pid=61907) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4910}} [ 1132.104077] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.104482] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.104768] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.105039] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.105282] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.107573] env[61907]: INFO nova.compute.manager [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Terminating instance [ 1132.248069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.248188] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.248356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.248547] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.248720] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.250798] env[61907]: INFO nova.compute.manager [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Terminating instance [ 1132.611572] env[61907]: DEBUG nova.compute.manager [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1132.611752] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.612806] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e834f786-fd05-43c1-830c-5d7c1cb54ae1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.620791] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.621099] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec82a85b-1380-41c0-82d6-e5dc711c3e58 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.629683] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1132.629683] env[61907]: value = "task-1244462" [ 1132.629683] env[61907]: _type = "Task" [ 1132.629683] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.636711] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244462, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.754340] env[61907]: DEBUG nova.compute.manager [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1132.755227] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1132.756319] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6257b8f4-16a5-4a36-ae22-0371ee58e28c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.764256] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1132.764573] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17fc00a2-0ac6-4c67-a548-b72ba1e31061 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.770596] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1132.770596] env[61907]: value = "task-1244463" [ 1132.770596] env[61907]: _type = "Task" [ 1132.770596] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.779798] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244463, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.129945] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1133.130304] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1133.130439] env[61907]: DEBUG nova.compute.manager [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1133.131344] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c78a6d6-11b9-44b7-82ce-a7167899cf0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.142578] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244462, 'name': PowerOffVM_Task, 'duration_secs': 0.189505} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.143985] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.144185] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1133.144474] env[61907]: DEBUG nova.compute.manager [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1133.145038] env[61907]: DEBUG nova.objects.instance [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.146770] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff03926d-9c0a-438f-ab42-12123012f618 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.230557] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.230790] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.230979] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] 0e9c2580-cf31-45fd-b5a7-ee13291056aa {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.231268] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b0f7927-8b22-402e-885b-4a5cfad77077 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.237984] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1133.237984] env[61907]: value = "task-1244465" [ 1133.237984] env[61907]: _type = "Task" [ 1133.237984] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.245661] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244465, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.278492] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244463, 'name': PowerOffVM_Task, 'duration_secs': 0.204968} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.278738] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1133.278941] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1133.279189] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79ae9a7f-0352-47d3-98d7-ef776a69706a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.347297] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1133.347528] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1133.347713] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleting the datastore file [datastore2] 378b5c99-5dbc-4657-8504-58d20dc6c9ea {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.347972] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aba6c013-8a6e-4e52-bb75-dcc0c77c5ae9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.354435] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1133.354435] env[61907]: value = "task-1244467" [ 1133.354435] env[61907]: _type = "Task" [ 1133.354435] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.361493] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244467, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.747966] env[61907]: DEBUG oslo_vmware.api [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244465, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131424} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.747966] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.748143] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.748240] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.748423] env[61907]: INFO nova.compute.manager [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1133.748672] env[61907]: DEBUG oslo.service.loopingcall [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.748872] env[61907]: DEBUG nova.compute.manager [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1133.748969] env[61907]: DEBUG nova.network.neutron [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1133.864487] env[61907]: DEBUG oslo_vmware.api [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244467, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158962} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.864763] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.864953] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1133.865153] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1133.865335] env[61907]: INFO nova.compute.manager [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1133.865774] env[61907]: DEBUG oslo.service.loopingcall [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.865774] env[61907]: DEBUG nova.compute.manager [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1133.865878] env[61907]: DEBUG nova.network.neutron [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1134.064798] env[61907]: DEBUG nova.compute.manager [req-87ce5bdb-7703-4fb7-be37-71a9e5f1f320 req-330095c2-d1c6-4fe4-8cb1-bc2e95f56217 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Received event network-vif-deleted-ce2db7f4-56a8-43ee-904e-1e7c22aa7512 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1134.064976] env[61907]: INFO nova.compute.manager [req-87ce5bdb-7703-4fb7-be37-71a9e5f1f320 req-330095c2-d1c6-4fe4-8cb1-bc2e95f56217 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Neutron deleted interface ce2db7f4-56a8-43ee-904e-1e7c22aa7512; detaching it from the instance and deleting it from the info cache [ 1134.065173] env[61907]: DEBUG nova.network.neutron [req-87ce5bdb-7703-4fb7-be37-71a9e5f1f320 req-330095c2-d1c6-4fe4-8cb1-bc2e95f56217 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.154337] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.154742] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf4eda72-7e52-4ca3-a0c2-3856f86e8934 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.161962] env[61907]: DEBUG oslo_vmware.api [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1134.161962] env[61907]: value = "task-1244468" [ 1134.161962] env[61907]: _type = "Task" [ 1134.161962] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.169842] env[61907]: DEBUG oslo_vmware.api [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244468, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.512058] env[61907]: DEBUG nova.network.neutron [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.565466] env[61907]: DEBUG nova.compute.manager [req-c79051dc-aa30-4e89-849d-23ee75a276f1 req-b4bceaae-a9a9-4019-b785-398a5fd71219 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Received event network-vif-deleted-9fa8da7e-09f8-42c1-8705-6a29e9024ad5 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1134.565812] env[61907]: INFO nova.compute.manager [req-c79051dc-aa30-4e89-849d-23ee75a276f1 req-b4bceaae-a9a9-4019-b785-398a5fd71219 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Neutron deleted interface 9fa8da7e-09f8-42c1-8705-6a29e9024ad5; detaching it from the instance and deleting it from the info cache [ 1134.566081] env[61907]: DEBUG nova.network.neutron [req-c79051dc-aa30-4e89-849d-23ee75a276f1 req-b4bceaae-a9a9-4019-b785-398a5fd71219 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.567994] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a65a036f-177c-4f50-a4f7-59c7a7610a65 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.578811] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5d5ee3-56b4-46dc-badb-7ec160973c42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.602305] env[61907]: DEBUG nova.compute.manager [req-87ce5bdb-7703-4fb7-be37-71a9e5f1f320 req-330095c2-d1c6-4fe4-8cb1-bc2e95f56217 service nova] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Detach interface failed, port_id=ce2db7f4-56a8-43ee-904e-1e7c22aa7512, reason: Instance 0e9c2580-cf31-45fd-b5a7-ee13291056aa could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1134.670499] env[61907]: DEBUG oslo_vmware.api [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244468, 'name': PowerOffVM_Task, 'duration_secs': 0.218364} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.671051] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1134.671051] env[61907]: DEBUG nova.compute.manager [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1134.671730] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65af4261-9f6f-47e5-a71f-aab0a34cfa80 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.014344] env[61907]: INFO nova.compute.manager [-] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Took 1.27 seconds to deallocate network for instance. [ 1135.043561] env[61907]: DEBUG nova.network.neutron [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.069121] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fccb10e5-802f-4ccd-9c2d-9a5678e7ef26 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.080422] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5087584f-9ad5-48b4-8f9d-7eef5c04ef08 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.105674] env[61907]: DEBUG nova.compute.manager [req-c79051dc-aa30-4e89-849d-23ee75a276f1 req-b4bceaae-a9a9-4019-b785-398a5fd71219 service nova] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Detach interface failed, port_id=9fa8da7e-09f8-42c1-8705-6a29e9024ad5, reason: Instance 378b5c99-5dbc-4657-8504-58d20dc6c9ea could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1135.183525] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8fa055b-d48b-4355-ae15-144f167aafad tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.053s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.521066] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.521362] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.521594] env[61907]: DEBUG nova.objects.instance [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid 0e9c2580-cf31-45fd-b5a7-ee13291056aa {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.546612] env[61907]: INFO nova.compute.manager [-] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Took 1.68 seconds to deallocate network for instance. [ 1136.054697] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.101657] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2357c9c1-31ee-4927-93cb-c07a9cd8fe8b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.109088] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12564334-f1e9-43f8-91a1-60d80b1b0bbe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.139913] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadb8608-125f-4b34-8b5e-0e5a90021dd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.147061] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece23bab-148c-4ca7-bbd8-e9119c85a01b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.159855] env[61907]: DEBUG nova.compute.provider_tree [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.285537] env[61907]: DEBUG nova.compute.manager [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Stashing vm_state: stopped {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1136.662817] env[61907]: DEBUG nova.scheduler.client.report [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1136.803561] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.168332] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.172015] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.116s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.172327] env[61907]: DEBUG nova.objects.instance [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'resources' on Instance uuid 378b5c99-5dbc-4657-8504-58d20dc6c9ea {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.186480] env[61907]: INFO nova.scheduler.client.report [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance 0e9c2580-cf31-45fd-b5a7-ee13291056aa [ 1137.694955] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f6a5a9c3-26b7-44ae-ba39-b6c8e5c62c5b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "0e9c2580-cf31-45fd-b5a7-ee13291056aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.590s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.744105] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed66613-24e1-4a19-880f-d562ab818479 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.752063] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada88ed1-3e01-43a4-97f7-351d11969e33 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.784859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82016475-8f70-41c1-8ace-08bb4694b31a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.791324] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6e7b3b-6136-492e-a409-928f8a8ec6a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.803924] env[61907]: DEBUG nova.compute.provider_tree [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.306934] env[61907]: DEBUG nova.scheduler.client.report [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1138.812131] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.641s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.814402] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.011s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.829294] env[61907]: INFO nova.scheduler.client.report [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted allocations for instance 378b5c99-5dbc-4657-8504-58d20dc6c9ea [ 1139.171044] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "868a8e82-db37-4a6d-86e7-64811c3c4840" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.171288] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.318441] env[61907]: INFO nova.compute.claims [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1139.336737] env[61907]: DEBUG oslo_concurrency.lockutils [None req-26e97254-25e9-4762-bee6-32087a896fa2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "378b5c99-5dbc-4657-8504-58d20dc6c9ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.088s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.673901] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1139.824264] env[61907]: INFO nova.compute.resource_tracker [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating resource usage from migration 57ca5678-80d7-424e-ad55-62029c854dbd [ 1139.889065] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6b5d53-1d3c-4e05-88f0-9d8a8c86a7b8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.896759] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b355aa76-457a-430c-99c1-d82f5b7b3551 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.575063] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bb81db-4bd7-44ca-a5e8-c83d923d074c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.578200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.578419] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.584847] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff3389f8-d652-468d-a555-3d67e039c2a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.589183] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.599089] env[61907]: DEBUG nova.compute.provider_tree [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1141.081475] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1141.103150] env[61907]: DEBUG nova.scheduler.client.report [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1141.600035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.607903] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.793s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.608124] env[61907]: INFO nova.compute.manager [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Migrating [ 1141.614402] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.025s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.615825] env[61907]: INFO nova.compute.claims [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.125562] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.125922] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.125922] env[61907]: DEBUG nova.network.neutron [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1142.701271] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dba44fc-60eb-4aa0-bc0e-791be39cb090 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.709178] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcdba2d-126b-4c50-9933-f2a4c2fc67ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.740626] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1093f5af-5da5-4b4f-9df2-350b5caa11c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.747648] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179c7a38-04d2-48bd-ae37-9d083629b492 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.760780] env[61907]: DEBUG nova.compute.provider_tree [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.854802] env[61907]: DEBUG nova.network.neutron [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.266100] env[61907]: DEBUG nova.scheduler.client.report [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1143.357611] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.770636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.771410] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1143.777248] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.177s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.778717] env[61907]: INFO nova.compute.claims [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1144.279300] env[61907]: DEBUG nova.compute.utils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.281518] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1144.281718] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1144.353858] env[61907]: DEBUG nova.policy [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1144.608174] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Successfully created port: ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1144.784763] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1144.872968] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df40b41b-c8f6-40e3-b42e-1cdf06828741 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.877048] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c618c76-3a23-4fdf-85d6-4ab580f74071 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.896386] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb98f9b-ec00-4d3d-a2f7-aa35af079048 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.899596] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1144.930112] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63e1437-276b-4d67-8201-97da5b2420bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.937543] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513ee8f2-4aa4-4ff2-9b9e-d10fc1bda8df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.952316] env[61907]: DEBUG nova.compute.provider_tree [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.405698] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1145.406039] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-820b271b-25c5-43dd-b91f-4d2dfb519d16 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.414300] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1145.414300] env[61907]: value = "task-1244469" [ 1145.414300] env[61907]: _type = "Task" [ 1145.414300] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.421924] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244469, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.455416] env[61907]: DEBUG nova.scheduler.client.report [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1145.796975] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1145.821691] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.821945] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.822127] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.822320] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.822473] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.822627] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.822840] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.823009] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.823195] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.823366] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.823545] env[61907]: DEBUG nova.virt.hardware [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.824441] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be411a1a-d5d8-4a63-b85f-ded3ecf58161 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.832446] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a303be-1dc7-448e-b4f4-245e22705587 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.923397] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1145.923598] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1145.960807] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.183s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.960807] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1145.983532] env[61907]: DEBUG nova.compute.manager [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Received event network-vif-plugged-ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1145.983765] env[61907]: DEBUG oslo_concurrency.lockutils [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] Acquiring lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.983979] env[61907]: DEBUG oslo_concurrency.lockutils [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.984238] env[61907]: DEBUG oslo_concurrency.lockutils [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.984489] env[61907]: DEBUG nova.compute.manager [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] No waiting events found dispatching network-vif-plugged-ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1145.984613] env[61907]: WARNING nova.compute.manager [req-13ce5d3c-79e0-41ec-9ec6-e608f4b6d8b3 req-134f2f85-bc6b-490f-a9ce-1231fc127c4c service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Received unexpected event network-vif-plugged-ac3b6d98-31e1-450b-892a-b537d9ff3349 for instance with vm_state building and task_state spawning. [ 1146.077360] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Successfully updated port: ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1146.429648] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1146.430079] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1146.430079] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1146.430274] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1146.430425] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1146.430579] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1146.430790] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1146.430954] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1146.431147] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1146.431317] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1146.431497] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1146.436536] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-78a26a2f-0714-4abc-9578-2232478ad213 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.451889] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1146.451889] env[61907]: value = "task-1244470" [ 1146.451889] env[61907]: _type = "Task" [ 1146.451889] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.460933] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.466146] env[61907]: DEBUG nova.compute.utils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1146.467503] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1146.467690] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1146.541068] env[61907]: DEBUG nova.policy [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88b29a917e03445281f16fba6d536280', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7652e98cde994af28b7bac0b81547474', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1146.581569] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.581701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.581806] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1146.821928] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Successfully created port: 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1146.962811] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244470, 'name': ReconfigVM_Task, 'duration_secs': 0.144135} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.963147] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.971999] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1147.113554] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1147.295406] env[61907]: DEBUG nova.network.neutron [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Updating instance_info_cache with network_info: [{"id": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "address": "fa:16:3e:98:a7:ca", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac3b6d98-31", "ovs_interfaceid": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.469578] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.469901] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.469991] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.470198] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.470350] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.470505] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.470714] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.470878] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.471058] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.471230] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.471408] env[61907]: DEBUG nova.virt.hardware [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.476890] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1147.480280] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21be56a1-e01b-4d5c-a1df-ad4d22dc94ba {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.499338] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1147.499338] env[61907]: value = "task-1244471" [ 1147.499338] env[61907]: _type = "Task" [ 1147.499338] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.506946] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.798116] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.798417] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Instance network_info: |[{"id": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "address": "fa:16:3e:98:a7:ca", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac3b6d98-31", "ovs_interfaceid": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1147.798965] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:a7:ca', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ac3b6d98-31e1-450b-892a-b537d9ff3349', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.806296] env[61907]: DEBUG oslo.service.loopingcall [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.806576] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1147.806878] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6145eab-6bd0-4380-92cd-2b51eb83b103 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.826057] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.826057] env[61907]: value = "task-1244472" [ 1147.826057] env[61907]: _type = "Task" [ 1147.826057] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.833239] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244472, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.994840] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1148.009377] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244471, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.011549] env[61907]: DEBUG nova.compute.manager [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Received event network-changed-ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1148.011749] env[61907]: DEBUG nova.compute.manager [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Refreshing instance network info cache due to event network-changed-ac3b6d98-31e1-450b-892a-b537d9ff3349. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1148.011969] env[61907]: DEBUG oslo_concurrency.lockutils [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] Acquiring lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.012136] env[61907]: DEBUG oslo_concurrency.lockutils [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] Acquired lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.012308] env[61907]: DEBUG nova.network.neutron [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Refreshing network info cache for port ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1148.021101] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1148.021344] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1148.021508] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1148.021696] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1148.021852] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1148.022015] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1148.022386] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1148.022565] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1148.022741] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1148.022911] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1148.023103] env[61907]: DEBUG nova.virt.hardware [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1148.024142] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48adb675-76db-425d-9581-60b3515f7c57 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.032374] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0c5f13b-22b8-49d0-a9b3-ed2611e60e59 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.286807] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Successfully updated port: 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1148.335367] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244472, 'name': CreateVM_Task} progress is 25%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.510993] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244471, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.716732] env[61907]: DEBUG nova.network.neutron [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Updated VIF entry in instance network info cache for port ac3b6d98-31e1-450b-892a-b537d9ff3349. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1148.717164] env[61907]: DEBUG nova.network.neutron [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Updating instance_info_cache with network_info: [{"id": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "address": "fa:16:3e:98:a7:ca", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapac3b6d98-31", "ovs_interfaceid": "ac3b6d98-31e1-450b-892a-b537d9ff3349", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.789136] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.789307] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.789473] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1148.837661] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244472, 'name': CreateVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.012979] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244471, 'name': ReconfigVM_Task, 'duration_secs': 1.184422} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.013298] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1149.014103] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec7307c-3662-4686-b1d8-8a405214e0d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.036246] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1149.036593] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf85a121-b073-45ea-a260-ac0be93b4031 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.055542] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1149.055542] env[61907]: value = "task-1244473" [ 1149.055542] env[61907]: _type = "Task" [ 1149.055542] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.063907] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244473, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.219983] env[61907]: DEBUG oslo_concurrency.lockutils [req-7fad8b8e-1d3f-4435-96a1-8d476e72aa80 req-9ce07658-840c-49bc-9e76-bda34845e64f service nova] Releasing lock "refresh_cache-868a8e82-db37-4a6d-86e7-64811c3c4840" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.324438] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1149.336265] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244472, 'name': CreateVM_Task, 'duration_secs': 1.086029} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.336455] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1149.337223] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.337413] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.337763] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1149.338034] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f48410e8-c960-4ebe-86b0-2eda28ddd576 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.343774] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1149.343774] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526084f7-f66c-edb9-740b-23291fc0be93" [ 1149.343774] env[61907]: _type = "Task" [ 1149.343774] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.352148] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526084f7-f66c-edb9-740b-23291fc0be93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.478623] env[61907]: DEBUG nova.network.neutron [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.566369] env[61907]: DEBUG oslo_vmware.api [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244473, 'name': ReconfigVM_Task, 'duration_secs': 0.24922} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.566797] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Reconfigured VM instance instance-00000063 to attach disk [datastore1] f7959e23-8879-4285-93da-804dd3115f65/f7959e23-8879-4285-93da-804dd3115f65.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1149.567073] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1149.855863] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]526084f7-f66c-edb9-740b-23291fc0be93, 'name': SearchDatastore_Task, 'duration_secs': 0.009628} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.856197] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.856438] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1149.856722] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.856881] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.857077] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1149.857360] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c59df9e-2b0e-40c5-bdae-3b75dfbc509d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.866062] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1149.866259] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1149.867304] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69fe67e0-0f2c-4105-9034-4c2e1e4ff2ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.872663] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1149.872663] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5246b74c-de4b-e536-e5e1-b0ac3530491d" [ 1149.872663] env[61907]: _type = "Task" [ 1149.872663] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.880286] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5246b74c-de4b-e536-e5e1-b0ac3530491d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.981243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.981668] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Instance network_info: |[{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1149.982214] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:3f:fb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e2153f70-3d14-42ab-8bb3-be78296dd3b8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1a805adb-8e64-4323-8acf-fa4e9fcf7fdf', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1149.990976] env[61907]: DEBUG oslo.service.loopingcall [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1149.991237] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1149.991467] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6fbc948-55e2-4492-90a2-45f0a3e6099b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.010985] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1150.010985] env[61907]: value = "task-1244474" [ 1150.010985] env[61907]: _type = "Task" [ 1150.010985] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.019333] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244474, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.037465] env[61907]: DEBUG nova.compute.manager [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Received event network-vif-plugged-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1150.037758] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1150.037938] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1150.038137] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.038318] env[61907]: DEBUG nova.compute.manager [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] No waiting events found dispatching network-vif-plugged-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1150.038490] env[61907]: WARNING nova.compute.manager [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Received unexpected event network-vif-plugged-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf for instance with vm_state building and task_state spawning. [ 1150.038656] env[61907]: DEBUG nova.compute.manager [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Received event network-changed-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1150.038834] env[61907]: DEBUG nova.compute.manager [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Refreshing instance network info cache due to event network-changed-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1150.039054] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Acquiring lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.039201] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Acquired lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.039366] env[61907]: DEBUG nova.network.neutron [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Refreshing network info cache for port 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1150.074352] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e26161-1243-4d55-a5ad-b332fe905152 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.093889] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e4d1dcb-14a6-4b08-aa05-e89f526daa9e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.111794] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1150.382706] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5246b74c-de4b-e536-e5e1-b0ac3530491d, 'name': SearchDatastore_Task, 'duration_secs': 0.009658} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.383490] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9b1954f-3b1d-46c2-9ee4-74a6308d7ca9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.388152] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1150.388152] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9351-3094-0799-bc10-d79feca1073b" [ 1150.388152] env[61907]: _type = "Task" [ 1150.388152] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.395318] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9351-3094-0799-bc10-d79feca1073b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.520243] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244474, 'name': CreateVM_Task, 'duration_secs': 0.318383} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.520426] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1150.521129] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.521312] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.521655] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1150.521919] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc2bd0e3-2c2c-408f-aa0a-952740d3fc97 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.526606] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1150.526606] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c1cd9-c0d4-fbfb-a255-a133dff48273" [ 1150.526606] env[61907]: _type = "Task" [ 1150.526606] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.535998] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c1cd9-c0d4-fbfb-a255-a133dff48273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.650318] env[61907]: DEBUG nova.network.neutron [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Port 6e35dd41-7b4c-44e2-8076-321b9b9bfc61 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1150.763880] env[61907]: DEBUG nova.network.neutron [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updated VIF entry in instance network info cache for port 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1150.764267] env[61907]: DEBUG nova.network.neutron [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.898539] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522c9351-3094-0799-bc10-d79feca1073b, 'name': SearchDatastore_Task, 'duration_secs': 0.008258} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.898801] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.899072] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 868a8e82-db37-4a6d-86e7-64811c3c4840/868a8e82-db37-4a6d-86e7-64811c3c4840.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1150.899341] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e4b58e1d-bfde-48fb-952b-8ddbde8856a6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.906137] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1150.906137] env[61907]: value = "task-1244475" [ 1150.906137] env[61907]: _type = "Task" [ 1150.906137] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.913560] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244475, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.038006] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529c1cd9-c0d4-fbfb-a255-a133dff48273, 'name': SearchDatastore_Task, 'duration_secs': 0.008329} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.038408] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.038688] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.038982] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.039200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.039438] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.039758] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4e114855-338f-47a7-aad6-7e99ee38b23c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.048303] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.048501] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1151.049307] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dabc282-b340-409a-90d4-27f22a42159d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.057882] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1151.057882] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5256127b-1e70-2b26-c57c-6910f361cb69" [ 1151.057882] env[61907]: _type = "Task" [ 1151.057882] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.066468] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5256127b-1e70-2b26-c57c-6910f361cb69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.268298] env[61907]: DEBUG oslo_concurrency.lockutils [req-6b94514d-081d-4c4f-b745-25431aff0ed9 req-4246bcac-abff-4c1c-9760-efdcf9af1a40 service nova] Releasing lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.415947] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244475, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.425253} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.416258] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 868a8e82-db37-4a6d-86e7-64811c3c4840/868a8e82-db37-4a6d-86e7-64811c3c4840.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1151.416476] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1151.416850] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-19c82e50-4e24-415e-800d-15477e918a3a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.422627] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1151.422627] env[61907]: value = "task-1244476" [ 1151.422627] env[61907]: _type = "Task" [ 1151.422627] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.429574] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244476, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.567215] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5256127b-1e70-2b26-c57c-6910f361cb69, 'name': SearchDatastore_Task, 'duration_secs': 0.060329} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.567959] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3de67d47-ba96-4781-8a2f-6ca90e4f7cb3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.572452] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1151.572452] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52afccc6-fcd5-8e1a-b7bb-b4278c92b484" [ 1151.572452] env[61907]: _type = "Task" [ 1151.572452] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.579467] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52afccc6-fcd5-8e1a-b7bb-b4278c92b484, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.674369] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.674789] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.674789] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.932250] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244476, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058877} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.932481] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1151.933238] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22b780d9-798a-4aec-b5c6-12403dcc3e7f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.954403] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] 868a8e82-db37-4a6d-86e7-64811c3c4840/868a8e82-db37-4a6d-86e7-64811c3c4840.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.954917] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74b0f484-f578-4671-bcb1-1ce90554a046 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.972932] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1151.972932] env[61907]: value = "task-1244477" [ 1151.972932] env[61907]: _type = "Task" [ 1151.972932] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.980020] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244477, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.082459] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52afccc6-fcd5-8e1a-b7bb-b4278c92b484, 'name': SearchDatastore_Task, 'duration_secs': 0.008271} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.082746] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.083063] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f1dc869a-473d-4d7e-bf4b-c46498c55c56/f1dc869a-473d-4d7e-bf4b-c46498c55c56.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1152.083339] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad980fc6-2b75-4ac9-b571-484fa6e33568 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.089834] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1152.089834] env[61907]: value = "task-1244478" [ 1152.089834] env[61907]: _type = "Task" [ 1152.089834] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.097469] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244478, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.486562] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244477, 'name': ReconfigVM_Task, 'duration_secs': 0.330941} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.486994] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Reconfigured VM instance instance-0000006b to attach disk [datastore2] 868a8e82-db37-4a6d-86e7-64811c3c4840/868a8e82-db37-4a6d-86e7-64811c3c4840.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1152.487893] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-53efa10d-0af7-4471-8bf8-f5e28b0a5a4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.497643] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1152.497643] env[61907]: value = "task-1244479" [ 1152.497643] env[61907]: _type = "Task" [ 1152.497643] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.508105] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244479, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.599991] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244478, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.708065] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1152.708290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1152.708460] env[61907]: DEBUG nova.network.neutron [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1153.007338] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244479, 'name': Rename_Task, 'duration_secs': 0.201098} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.007607] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1153.007851] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0222f2b1-7edf-41a2-a7ac-6970dbe049ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.013860] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1153.013860] env[61907]: value = "task-1244480" [ 1153.013860] env[61907]: _type = "Task" [ 1153.013860] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.020717] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244480, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.099841] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244478, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.524931} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.100173] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] f1dc869a-473d-4d7e-bf4b-c46498c55c56/f1dc869a-473d-4d7e-bf4b-c46498c55c56.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1153.100400] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1153.100659] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e399270b-5cde-4857-9b8b-2af2b2e9a207 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.106708] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1153.106708] env[61907]: value = "task-1244481" [ 1153.106708] env[61907]: _type = "Task" [ 1153.106708] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.113866] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244481, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.429468] env[61907]: DEBUG nova.network.neutron [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.455946] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_power_states {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1153.524415] env[61907]: DEBUG oslo_vmware.api [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244480, 'name': PowerOnVM_Task, 'duration_secs': 0.500658} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.524694] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1153.524898] env[61907]: INFO nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1153.525092] env[61907]: DEBUG nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1153.525838] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ab0379-88b0-4af6-b030-6dd1b8b51fd3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.617194] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244481, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064708} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.617456] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1153.618227] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1672a4-aa2e-4378-81c1-61fb9590224c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.639660] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] f1dc869a-473d-4d7e-bf4b-c46498c55c56/f1dc869a-473d-4d7e-bf4b-c46498c55c56.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1153.639897] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f80db2f7-25b8-4e38-ac26-f4844e9341f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.657555] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1153.657555] env[61907]: value = "task-1244482" [ 1153.657555] env[61907]: _type = "Task" [ 1153.657555] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.664576] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.932244] env[61907]: DEBUG oslo_concurrency.lockutils [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1153.960528] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Getting list of instances from cluster (obj){ [ 1153.960528] env[61907]: value = "domain-c8" [ 1153.960528] env[61907]: _type = "ClusterComputeResource" [ 1153.960528] env[61907]: } {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1153.961019] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad64dca-7ace-47db-ad42-ac1751b0896a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.974574] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Got total of 5 instances {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1153.974574] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1153.974574] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1153.974574] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid b2c5689e-f930-4548-b061-fd2974d44575 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1153.974574] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid 868a8e82-db37-4a6d-86e7-64811c3c4840 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1153.974838] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1153.974956] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.975195] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.975532] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.975659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f7959e23-8879-4285-93da-804dd3115f65" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.975810] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f7959e23-8879-4285-93da-804dd3115f65] During sync_power_state the instance has a pending task (resize_migrated). Skip. [ 1153.975982] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f7959e23-8879-4285-93da-804dd3115f65" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.976212] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "b2c5689e-f930-4548-b061-fd2974d44575" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.976399] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "b2c5689e-f930-4548-b061-fd2974d44575" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.976670] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "868a8e82-db37-4a6d-86e7-64811c3c4840" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.977059] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.978409] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-220f4d7d-ebed-45f5-847d-a62b02e5ab2c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.981614] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0a223dd-89a0-42f3-aa6c-b9098268a229 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.042940] env[61907]: INFO nova.compute.manager [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Took 13.47 seconds to build instance. [ 1154.167528] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244482, 'name': ReconfigVM_Task, 'duration_secs': 0.266789} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.167888] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Reconfigured VM instance instance-0000006c to attach disk [datastore2] f1dc869a-473d-4d7e-bf4b-c46498c55c56/f1dc869a-473d-4d7e-bf4b-c46498c55c56.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1154.168539] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9113b36-4514-4f43-8c2a-33caa0842915 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.174428] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1154.174428] env[61907]: value = "task-1244483" [ 1154.174428] env[61907]: _type = "Task" [ 1154.174428] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.182410] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244483, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.455904] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ae6e9e-56e4-466c-b4e2-a16cd1af04a0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.474633] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fd6cac-8703-46e3-b22f-0095377cf71d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.481130] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1154.492354] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.517s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.492686] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "b2c5689e-f930-4548-b061-fd2974d44575" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.516s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.544643] env[61907]: DEBUG oslo_concurrency.lockutils [None req-15f7d894-52d1-46b9-a3ce-6f40075ab87d tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.373s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.544914] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.568s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.545133] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] During sync_power_state the instance has a pending task (spawning). Skip. [ 1154.545314] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.684084] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244483, 'name': Rename_Task, 'duration_secs': 0.14111} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.684319] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1154.684574] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bdd7f904-8c68-4016-b9c5-45e4b2cf88d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.690257] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1154.690257] env[61907]: value = "task-1244484" [ 1154.690257] env[61907]: _type = "Task" [ 1154.690257] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.705173] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.793694] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "868a8e82-db37-4a6d-86e7-64811c3c4840" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.794128] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.794394] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.794625] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.794808] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.797191] env[61907]: INFO nova.compute.manager [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Terminating instance [ 1154.987189] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-be66a09b-91a1-443b-94c2-41c53f476f3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance 'f7959e23-8879-4285-93da-804dd3115f65' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1155.200859] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244484, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.301189] env[61907]: DEBUG nova.compute.manager [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1155.301402] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1155.302241] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20d0de6-adbe-4dbe-83b5-ef16a7abd7e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.309498] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1155.309752] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9135c458-16cc-4b08-b65e-c5baddaa6fac {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.314980] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1155.314980] env[61907]: value = "task-1244485" [ 1155.314980] env[61907]: _type = "Task" [ 1155.314980] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.322319] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244485, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.700868] env[61907]: DEBUG oslo_vmware.api [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244484, 'name': PowerOnVM_Task, 'duration_secs': 0.858264} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.701157] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1155.701367] env[61907]: INFO nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Took 7.71 seconds to spawn the instance on the hypervisor. [ 1155.701551] env[61907]: DEBUG nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1155.702311] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-931572d3-b048-45f8-b54b-8c99791c46cc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.826171] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244485, 'name': PowerOffVM_Task, 'duration_secs': 0.188301} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.826397] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1155.826600] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1155.826854] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-970e1f1f-3914-4e0a-9afa-fdadb9472876 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.888543] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1155.888810] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1155.888950] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] 868a8e82-db37-4a6d-86e7-64811c3c4840 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1155.889224] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b62a4c1e-19f3-4add-bda1-4365d826f3f7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.895135] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1155.895135] env[61907]: value = "task-1244487" [ 1155.895135] env[61907]: _type = "Task" [ 1155.895135] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1155.902507] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244487, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1156.028695] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.029070] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1156.220930] env[61907]: INFO nova.compute.manager [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Took 14.63 seconds to build instance. [ 1156.405030] env[61907]: DEBUG oslo_vmware.api [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244487, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140631} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1156.405257] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1156.405451] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1156.405631] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1156.405816] env[61907]: INFO nova.compute.manager [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1156.406075] env[61907]: DEBUG oslo.service.loopingcall [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1156.406278] env[61907]: DEBUG nova.compute.manager [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1156.406376] env[61907]: DEBUG nova.network.neutron [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1156.724248] env[61907]: DEBUG nova.compute.manager [req-3bf897cc-1807-4288-b4e6-1e30ea647e54 req-7bac0597-55d0-40af-9ddc-2c772114e1b5 service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Received event network-vif-deleted-ac3b6d98-31e1-450b-892a-b537d9ff3349 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1156.724248] env[61907]: INFO nova.compute.manager [req-3bf897cc-1807-4288-b4e6-1e30ea647e54 req-7bac0597-55d0-40af-9ddc-2c772114e1b5 service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Neutron deleted interface ac3b6d98-31e1-450b-892a-b537d9ff3349; detaching it from the instance and deleting it from the info cache [ 1156.724248] env[61907]: DEBUG nova.network.neutron [req-3bf897cc-1807-4288-b4e6-1e30ea647e54 req-7bac0597-55d0-40af-9ddc-2c772114e1b5 service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.726793] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b208ffa1-038d-4229-bc67-09247850f0c2 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.148s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.729437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.752s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1156.729787] env[61907]: INFO nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] During sync_power_state the instance has a pending task (spawning). Skip. [ 1156.729850] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.838748] env[61907]: DEBUG nova.compute.manager [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Received event network-changed-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1156.838972] env[61907]: DEBUG nova.compute.manager [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Refreshing instance network info cache due to event network-changed-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1156.839211] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] Acquiring lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.839363] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] Acquired lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.839531] env[61907]: DEBUG nova.network.neutron [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Refreshing network info cache for port 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1157.164437] env[61907]: DEBUG nova.network.neutron [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.228351] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06c3d256-73cd-4ce5-a3c6-ac22acc467c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.238028] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdd4a884-d35f-462f-8032-f74f2d4b70e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.263134] env[61907]: DEBUG nova.compute.manager [req-3bf897cc-1807-4288-b4e6-1e30ea647e54 req-7bac0597-55d0-40af-9ddc-2c772114e1b5 service nova] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Detach interface failed, port_id=ac3b6d98-31e1-450b-892a-b537d9ff3349, reason: Instance 868a8e82-db37-4a6d-86e7-64811c3c4840 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1157.305359] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.305430] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.305596] env[61907]: DEBUG nova.compute.manager [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Going to confirm migration 6 {{(pid=61907) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5126}} [ 1157.513348] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1157.513521] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1157.553746] env[61907]: DEBUG nova.network.neutron [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updated VIF entry in instance network info cache for port 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1157.554146] env[61907]: DEBUG nova.network.neutron [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1157.667427] env[61907]: INFO nova.compute.manager [-] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Took 1.26 seconds to deallocate network for instance. [ 1157.845439] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1157.845717] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1157.845951] env[61907]: DEBUG nova.network.neutron [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1157.846309] env[61907]: DEBUG nova.objects.instance [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'info_cache' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.043484] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1158.056316] env[61907]: DEBUG oslo_concurrency.lockutils [req-aaa76637-ce17-4b6f-ba23-2a916665c26d req-059ebe51-94e5-4257-8ac8-578916798464 service nova] Releasing lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1158.173895] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.174250] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.174437] env[61907]: DEBUG nova.objects.instance [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid 868a8e82-db37-4a6d-86e7-64811c3c4840 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.759724] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7b2a281-de08-4f80-b064-30272911ae68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.766669] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a48261-c84b-4298-8c10-9cafe1c1693f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.795857] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd44e66-6732-4e02-be29-fda233c68f66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.802267] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6af20d-3727-4ff0-b63d-f2be6b9a9216 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.814649] env[61907]: DEBUG nova.compute.provider_tree [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.054601] env[61907]: DEBUG nova.network.neutron [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.317670] env[61907]: DEBUG nova.scheduler.client.report [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1159.557193] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1159.557480] env[61907]: DEBUG nova.objects.instance [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'migration_context' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1159.558632] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1159.559284] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f7959e23-8879-4285-93da-804dd3115f65] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1159.822667] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.648s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.847238] env[61907]: INFO nova.scheduler.client.report [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance 868a8e82-db37-4a6d-86e7-64811c3c4840 [ 1160.061598] env[61907]: DEBUG nova.objects.base [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1160.064543] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6bef354-3bf8-4f3e-8dc5-5077aa4a9fdb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.083446] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1322442e-da5b-4d10-b738-b052223acc20 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.088586] env[61907]: DEBUG oslo_vmware.api [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1160.088586] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f56e8f-db1f-2ae1-d724-2abd1812436c" [ 1160.088586] env[61907]: _type = "Task" [ 1160.088586] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.096058] env[61907]: DEBUG oslo_vmware.api [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f56e8f-db1f-2ae1-d724-2abd1812436c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.354784] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d5362abf-d980-4b84-a846-7073b9770c51 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "868a8e82-db37-4a6d-86e7-64811c3c4840" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.561s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.600984] env[61907]: DEBUG oslo_vmware.api [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52f56e8f-db1f-2ae1-d724-2abd1812436c, 'name': SearchDatastore_Task, 'duration_secs': 0.006627} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.601301] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.601529] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.789127] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1161.170859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7460ba2-be9f-4129-be9a-e6122e30c88a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.178804] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1213f16f-bdeb-47ae-b5f9-6106fecba1a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.208183] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e02fb12-e867-4082-a3cb-e434dd7d6924 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.215241] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3789e94e-ea9e-4ba8-be0d-7bda1cd07ec8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.227964] env[61907]: DEBUG nova.compute.provider_tree [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.293324] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1161.293324] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1161.293324] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.293324] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.293324] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.293324] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.293324] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1161.293324] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1161.730658] env[61907]: DEBUG nova.scheduler.client.report [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1161.774332] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.774585] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.795364] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.276359] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1162.740864] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.139s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.741373] env[61907]: DEBUG nova.compute.manager [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=61907) _confirm_resize /opt/stack/nova/nova/compute/manager.py:5252}} [ 1162.743941] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.949s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.744184] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.744350] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1162.745428] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa03a213-add8-482a-97b4-0b48a1c167d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.753896] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e272cc8-2263-4728-990c-be91cb4e0f96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.767751] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b3cca9c-3033-4b5d-a984-ef5c3cc1aa0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.774391] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92b5c59-6fab-429d-8d8a-b423bb5e7ada {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.805350] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180707MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1162.805547] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1162.805701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1162.823169] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.293400] env[61907]: INFO nova.scheduler.client.report [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted allocation for migration 57ca5678-80d7-424e-ad55-62029c854dbd [ 1163.799192] env[61907]: DEBUG oslo_concurrency.lockutils [None req-c8969f17-e31a-49c2-8627-d857cbf67d88 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.494s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.830232] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.830617] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance b2c5689e-f930-4548-b061-fd2974d44575 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.830617] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f7959e23-8879-4285-93da-804dd3115f65 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1163.830617] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f1dc869a-473d-4d7e-bf4b-c46498c55c56 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1164.245216] env[61907]: DEBUG nova.objects.instance [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.332863] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance ad8a7d32-6ef3-44d2-9617-49335af29ae2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1164.333137] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1164.333287] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1164.400126] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f127b909-b418-44c6-8d6b-cabbf72dcb76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.407715] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048a26ed-4598-49be-a5e6-8633b4dce014 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.437717] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e9fd322-b17e-4655-abd2-9bbf0194edb2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.445594] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85c10f1-b596-49d1-a2d5-2d24700a09ec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.458274] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1164.750636] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1164.750799] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1164.750985] env[61907]: DEBUG nova.network.neutron [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1164.751189] env[61907]: DEBUG nova.objects.instance [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'info_cache' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1164.961127] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1165.256177] env[61907]: DEBUG nova.objects.base [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1165.465562] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1165.465791] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.660s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1165.466097] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.643s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1165.467642] env[61907]: INFO nova.compute.claims [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1165.691277] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1165.960142] env[61907]: DEBUG nova.network.neutron [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [{"id": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "address": "fa:16:3e:05:03:28", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e35dd41-7b", "ovs_interfaceid": "6e35dd41-7b4c-44e2-8076-321b9b9bfc61", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1166.464032] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-f7959e23-8879-4285-93da-804dd3115f65" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1166.541387] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556f7d5c-2ff7-425b-a720-5c9d6c70b790 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.548810] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144154f8-05cd-4a4d-943e-6a511317a2e2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.577492] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0ebaa1-0648-4a82-a398-2f4f55dab34c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.583985] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a294a0ae-fd2d-445b-ad6b-4f88e52ea5ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.597349] env[61907]: DEBUG nova.compute.provider_tree [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1167.100370] env[61907]: DEBUG nova.scheduler.client.report [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1167.469339] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1167.469694] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05ec5601-0983-40e2-ba2b-bb722cac17ff {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1167.476807] env[61907]: DEBUG oslo_vmware.api [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1167.476807] env[61907]: value = "task-1244488" [ 1167.476807] env[61907]: _type = "Task" [ 1167.476807] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1167.484145] env[61907]: DEBUG oslo_vmware.api [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244488, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1167.605356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.139s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.605902] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1167.986648] env[61907]: DEBUG oslo_vmware.api [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244488, 'name': PowerOnVM_Task, 'duration_secs': 0.37015} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.986946] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1167.987182] env[61907]: DEBUG nova.compute.manager [None req-0127a3b7-fed9-41bf-8b6a-44143e16bbc3 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1167.987967] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e64c8dd-d2e0-463d-af59-e59144c30c42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.111731] env[61907]: DEBUG nova.compute.utils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1168.113037] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1168.113193] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1168.151093] env[61907]: DEBUG nova.policy [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b8123b43a5094cb8a8eaf665a7ff6bb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7581e815d75647c58ebe4198b7ed5c31', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1168.388686] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Successfully created port: e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1168.616787] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1168.812219] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.812467] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.812695] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "f7959e23-8879-4285-93da-804dd3115f65-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.812889] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.813094] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.815264] env[61907]: INFO nova.compute.manager [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Terminating instance [ 1169.318643] env[61907]: DEBUG nova.compute.manager [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1169.318880] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1169.319800] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae41fb7-76e3-410e-8884-822253bdd895 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.327898] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1169.328103] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a2e14e8d-2fcc-4844-8081-8b4982db4812 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.334277] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1169.334277] env[61907]: value = "task-1244489" [ 1169.334277] env[61907]: _type = "Task" [ 1169.334277] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.341965] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244489, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.626129] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1169.655823] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1169.656249] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1169.656519] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1169.656857] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1169.657140] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1169.657413] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1169.657758] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1169.658052] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1169.658346] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1169.658612] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1169.658898] env[61907]: DEBUG nova.virt.hardware [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1169.660136] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc717bb2-d567-4e98-8057-f4d3f93a6fa8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.670714] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5397c39a-d86f-4f2f-9c8b-09c4644a7990 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.844184] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244489, 'name': PowerOffVM_Task, 'duration_secs': 0.201187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1169.844430] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1169.844626] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1169.844865] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e52d6b49-91ea-4c93-83a3-0da8c28daff7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.877496] env[61907]: DEBUG nova.compute.manager [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Received event network-vif-plugged-e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1169.877718] env[61907]: DEBUG oslo_concurrency.lockutils [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] Acquiring lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.878310] env[61907]: DEBUG oslo_concurrency.lockutils [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.878504] env[61907]: DEBUG oslo_concurrency.lockutils [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.878686] env[61907]: DEBUG nova.compute.manager [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] No waiting events found dispatching network-vif-plugged-e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1169.878911] env[61907]: WARNING nova.compute.manager [req-60c7266e-8558-4d2a-be32-817751975710 req-032b22fa-5401-4957-b633-15129b81c525 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Received unexpected event network-vif-plugged-e21af03c-7e0e-4275-b1c9-a14a3accb004 for instance with vm_state building and task_state spawning. [ 1169.912411] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1169.912662] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1169.912835] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleting the datastore file [datastore1] f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1169.913115] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d42595fc-f4c1-43d7-998b-8ecbfa2062ce {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.919530] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1169.919530] env[61907]: value = "task-1244491" [ 1169.919530] env[61907]: _type = "Task" [ 1169.919530] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1169.926960] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244491, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1169.963045] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Successfully updated port: e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1170.429422] env[61907]: DEBUG oslo_vmware.api [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244491, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147981} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.429692] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1170.429882] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1170.430088] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1170.430273] env[61907]: INFO nova.compute.manager [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1170.430520] env[61907]: DEBUG oslo.service.loopingcall [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1170.430716] env[61907]: DEBUG nova.compute.manager [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1170.430814] env[61907]: DEBUG nova.network.neutron [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1170.466166] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1170.466268] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1170.466423] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1171.010978] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1171.147234] env[61907]: DEBUG nova.network.neutron [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Updating instance_info_cache with network_info: [{"id": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "address": "fa:16:3e:4e:24:3e", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21af03c-7e", "ovs_interfaceid": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.592316] env[61907]: DEBUG nova.network.neutron [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1171.650665] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1171.651032] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Instance network_info: |[{"id": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "address": "fa:16:3e:4e:24:3e", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21af03c-7e", "ovs_interfaceid": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1171.651494] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4e:24:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7b2049d7-f99e-425a-afdb-2c95ca88e483', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e21af03c-7e0e-4275-b1c9-a14a3accb004', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1171.660914] env[61907]: DEBUG oslo.service.loopingcall [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1171.661809] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1171.662064] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eba1204c-8ee4-451f-bc4f-d2cb69bc3c79 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1171.682520] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1171.682520] env[61907]: value = "task-1244492" [ 1171.682520] env[61907]: _type = "Task" [ 1171.682520] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1171.691962] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244492, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1171.911247] env[61907]: DEBUG nova.compute.manager [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Received event network-changed-e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1171.911452] env[61907]: DEBUG nova.compute.manager [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Refreshing instance network info cache due to event network-changed-e21af03c-7e0e-4275-b1c9-a14a3accb004. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1171.911675] env[61907]: DEBUG oslo_concurrency.lockutils [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] Acquiring lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1171.911826] env[61907]: DEBUG oslo_concurrency.lockutils [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] Acquired lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1171.912096] env[61907]: DEBUG nova.network.neutron [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Refreshing network info cache for port e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1172.094516] env[61907]: INFO nova.compute.manager [-] [instance: f7959e23-8879-4285-93da-804dd3115f65] Took 1.66 seconds to deallocate network for instance. [ 1172.192351] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244492, 'name': CreateVM_Task, 'duration_secs': 0.303751} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.192522] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1172.193281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.193480] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.193828] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1172.194118] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8cf077b7-28b5-422f-b398-726b7c7f2f71 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.200026] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1172.200026] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ac0974-aef9-91fd-4d40-841de20eb02a" [ 1172.200026] env[61907]: _type = "Task" [ 1172.200026] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.207214] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ac0974-aef9-91fd-4d40-841de20eb02a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1172.603894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1172.604193] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.604420] env[61907]: DEBUG nova.objects.instance [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'resources' on Instance uuid f7959e23-8879-4285-93da-804dd3115f65 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1172.610150] env[61907]: DEBUG nova.network.neutron [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Updated VIF entry in instance network info cache for port e21af03c-7e0e-4275-b1c9-a14a3accb004. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1172.610484] env[61907]: DEBUG nova.network.neutron [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Updating instance_info_cache with network_info: [{"id": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "address": "fa:16:3e:4e:24:3e", "network": {"id": "1284f882-32ee-4ac5-8a16-f0a015d75bfd", "bridge": "br-int", "label": "tempest-ServersTestJSON-2041147658-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7581e815d75647c58ebe4198b7ed5c31", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7b2049d7-f99e-425a-afdb-2c95ca88e483", "external-id": "nsx-vlan-transportzone-803", "segmentation_id": 803, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape21af03c-7e", "ovs_interfaceid": "e21af03c-7e0e-4275-b1c9-a14a3accb004", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1172.709974] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52ac0974-aef9-91fd-4d40-841de20eb02a, 'name': SearchDatastore_Task, 'duration_secs': 0.010841} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1172.710250] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1172.710492] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1172.710731] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1172.710881] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1172.711105] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1172.711392] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d645672c-4270-4b56-9aef-4c09dd985df6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.719795] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1172.720017] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1172.720716] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05d212ab-30d2-4110-b4c9-915c5aaf13e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1172.725588] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1172.725588] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52384d16-ecab-83d3-adbf-36aa084ce08a" [ 1172.725588] env[61907]: _type = "Task" [ 1172.725588] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1172.732791] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52384d16-ecab-83d3-adbf-36aa084ce08a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.113552] env[61907]: DEBUG oslo_concurrency.lockutils [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] Releasing lock "refresh_cache-ad8a7d32-6ef3-44d2-9617-49335af29ae2" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.113827] env[61907]: DEBUG nova.compute.manager [req-ab7929bc-96f4-4e61-a3b1-2cd8ec8ec686 req-e78319a0-b3dd-4cf8-a8c9-fdd96f854814 service nova] [instance: f7959e23-8879-4285-93da-804dd3115f65] Received event network-vif-deleted-6e35dd41-7b4c-44e2-8076-321b9b9bfc61 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1173.182523] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452f724e-5035-4989-82fb-56b6426b00dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.189992] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5de60ae-6e6a-47a1-b6ae-13e6bccb7d7b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.220362] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671e02c2-548a-423a-a156-001f9be739a4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.230159] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9fc2b4-5603-4196-afa7-482cd9971d9d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.238994] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52384d16-ecab-83d3-adbf-36aa084ce08a, 'name': SearchDatastore_Task, 'duration_secs': 0.009095} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.247039] env[61907]: DEBUG nova.compute.provider_tree [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.248187] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bda8a52-0c15-4da4-8133-e445424bda02 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.253327] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1173.253327] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5214d4bd-a8ae-d0f3-542d-28bee7ccb4b2" [ 1173.253327] env[61907]: _type = "Task" [ 1173.253327] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.261146] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5214d4bd-a8ae-d0f3-542d-28bee7ccb4b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.752336] env[61907]: DEBUG nova.scheduler.client.report [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1173.764842] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5214d4bd-a8ae-d0f3-542d-28bee7ccb4b2, 'name': SearchDatastore_Task, 'duration_secs': 0.016826} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.765666] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1173.765928] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ad8a7d32-6ef3-44d2-9617-49335af29ae2/ad8a7d32-6ef3-44d2-9617-49335af29ae2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1173.766206] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcf6c6eb-7612-4924-8074-8e2ac19375a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.772440] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1173.772440] env[61907]: value = "task-1244493" [ 1173.772440] env[61907]: _type = "Task" [ 1173.772440] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.780154] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.260437] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.656s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.277994] env[61907]: INFO nova.scheduler.client.report [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted allocations for instance f7959e23-8879-4285-93da-804dd3115f65 [ 1174.284559] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.453301} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.285016] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] ad8a7d32-6ef3-44d2-9617-49335af29ae2/ad8a7d32-6ef3-44d2-9617-49335af29ae2.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1174.285253] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1174.285653] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ec3a3b53-a5db-4431-a96d-a83bf05f604f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.292199] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1174.292199] env[61907]: value = "task-1244494" [ 1174.292199] env[61907]: _type = "Task" [ 1174.292199] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.299118] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244494, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.788228] env[61907]: DEBUG oslo_concurrency.lockutils [None req-80d29c12-ef9f-4435-8797-1531b21be87b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "f7959e23-8879-4285-93da-804dd3115f65" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.976s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.801340] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244494, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062719} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.801628] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1174.802387] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b00a36b-0b00-4e7d-a10a-71e7050f844e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.823826] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] ad8a7d32-6ef3-44d2-9617-49335af29ae2/ad8a7d32-6ef3-44d2-9617-49335af29ae2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1174.824276] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b4dc945-83e9-4fbc-a16c-4b502c286f96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.844293] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1174.844293] env[61907]: value = "task-1244495" [ 1174.844293] env[61907]: _type = "Task" [ 1174.844293] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1174.851468] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.355676] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244495, 'name': ReconfigVM_Task, 'duration_secs': 0.275761} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.356056] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Reconfigured VM instance instance-0000006d to attach disk [datastore2] ad8a7d32-6ef3-44d2-9617-49335af29ae2/ad8a7d32-6ef3-44d2-9617-49335af29ae2.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1175.356742] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3139fba-0920-48a3-b40b-53dda02055f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.362939] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1175.362939] env[61907]: value = "task-1244496" [ 1175.362939] env[61907]: _type = "Task" [ 1175.362939] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.370624] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244496, 'name': Rename_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.873196] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244496, 'name': Rename_Task, 'duration_secs': 0.147197} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1175.873555] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1175.873831] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f9fac348-792e-4ba0-9e38-5d32cd28a2c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1175.881051] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1175.881051] env[61907]: value = "task-1244497" [ 1175.881051] env[61907]: _type = "Task" [ 1175.881051] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1175.887738] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1175.964552] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1175.964821] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.390506] env[61907]: DEBUG oslo_vmware.api [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244497, 'name': PowerOnVM_Task, 'duration_secs': 0.408965} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1176.390984] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1176.391236] env[61907]: INFO nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Took 6.76 seconds to spawn the instance on the hypervisor. [ 1176.391502] env[61907]: DEBUG nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1176.392321] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ae161a-6310-480f-8bf9-1dc3f755f169 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.467511] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1176.908020] env[61907]: INFO nova.compute.manager [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Took 14.13 seconds to build instance. [ 1177.046007] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.046297] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.047897] env[61907]: INFO nova.compute.claims [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1177.383683] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1177.409736] env[61907]: DEBUG oslo_concurrency.lockutils [None req-921b9775-e4e9-4399-a5d5-381a6d719cd1 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.635s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.410067] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.027s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1177.410199] env[61907]: DEBUG nova.compute.manager [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1177.411162] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc98014-19f8-4ca7-9c5b-9a9b496fc6de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.418280] env[61907]: DEBUG nova.compute.manager [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1177.418877] env[61907]: DEBUG nova.objects.instance [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'flavor' on Instance uuid ad8a7d32-6ef3-44d2-9617-49335af29ae2 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1178.119056] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9705d90-e243-4b92-8ba7-ae49fb724af0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.126252] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dba20d-8add-4302-a744-fbb1a1a63551 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.154827] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99c717d1-6f42-4501-b45d-f2a5b6567f0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.161588] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52473b5f-bdbf-46c5-b94c-a32533754cd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.174011] env[61907]: DEBUG nova.compute.provider_tree [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1178.425684] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.426068] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d0d7f5f7-5c45-4d7b-b79a-424840a9e8d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.434912] env[61907]: DEBUG oslo_vmware.api [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1178.434912] env[61907]: value = "task-1244498" [ 1178.434912] env[61907]: _type = "Task" [ 1178.434912] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.442287] env[61907]: DEBUG oslo_vmware.api [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244498, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.677779] env[61907]: DEBUG nova.scheduler.client.report [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1178.945099] env[61907]: DEBUG oslo_vmware.api [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244498, 'name': PowerOffVM_Task, 'duration_secs': 0.175072} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.945413] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.945619] env[61907]: DEBUG nova.compute.manager [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1178.946400] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9045bdb3-170a-4e7a-b3d9-df06496c9057 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.183258] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.183943] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1179.458581] env[61907]: DEBUG oslo_concurrency.lockutils [None req-4e348597-397e-447d-aed5-1bdcc63f1f15 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.048s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1179.689367] env[61907]: DEBUG nova.compute.utils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1179.690774] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1179.690939] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1179.726294] env[61907]: DEBUG nova.policy [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8186074291264496984da491880a55b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f851fef407334ba2919877cfd4865435', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1179.991694] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Successfully created port: e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1180.021408] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.021653] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.021879] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1180.022086] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1180.022269] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1180.024513] env[61907]: INFO nova.compute.manager [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Terminating instance [ 1180.193636] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1180.528137] env[61907]: DEBUG nova.compute.manager [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1180.528402] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1180.529433] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ca07e6-5596-4e13-a80b-b9a34890b349 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.537394] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1180.537644] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a612ef6f-eb0c-422e-a872-4ad435ccd22c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.602193] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1180.602440] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1180.602633] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore2] ad8a7d32-6ef3-44d2-9617-49335af29ae2 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1180.602942] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca3b52bf-fb24-4856-8bfd-3769ba168eb6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.609522] env[61907]: DEBUG oslo_vmware.api [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1180.609522] env[61907]: value = "task-1244500" [ 1180.609522] env[61907]: _type = "Task" [ 1180.609522] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.618034] env[61907]: DEBUG oslo_vmware.api [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244500, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.120089] env[61907]: DEBUG oslo_vmware.api [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244500, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.153141} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.120357] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1181.120547] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1181.120724] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1181.120907] env[61907]: INFO nova.compute.manager [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1181.121176] env[61907]: DEBUG oslo.service.loopingcall [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1181.121372] env[61907]: DEBUG nova.compute.manager [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1181.121465] env[61907]: DEBUG nova.network.neutron [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1181.203309] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1181.240915] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1181.241198] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1181.241368] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1181.241563] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1181.241718] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1181.241876] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1181.242346] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1181.242600] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1181.242826] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1181.243055] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1181.243398] env[61907]: DEBUG nova.virt.hardware [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1181.244301] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da87ac1-d54d-42c4-af5c-fb67124ad4f1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.252681] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d90a5b-eb4c-4c1f-be31-8ac0eadf53e1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.364177] env[61907]: DEBUG nova.compute.manager [req-34bf3580-a8ad-45d6-90ba-aa02974ef04c req-205550b4-3864-45c5-ace8-fbe4a14380fb service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Received event network-vif-deleted-e21af03c-7e0e-4275-b1c9-a14a3accb004 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1181.364177] env[61907]: INFO nova.compute.manager [req-34bf3580-a8ad-45d6-90ba-aa02974ef04c req-205550b4-3864-45c5-ace8-fbe4a14380fb service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Neutron deleted interface e21af03c-7e0e-4275-b1c9-a14a3accb004; detaching it from the instance and deleting it from the info cache [ 1181.364177] env[61907]: DEBUG nova.network.neutron [req-34bf3580-a8ad-45d6-90ba-aa02974ef04c req-205550b4-3864-45c5-ace8-fbe4a14380fb service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.375665] env[61907]: DEBUG nova.compute.manager [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Received event network-vif-plugged-e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1181.376083] env[61907]: DEBUG oslo_concurrency.lockutils [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1181.376557] env[61907]: DEBUG oslo_concurrency.lockutils [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1181.376883] env[61907]: DEBUG oslo_concurrency.lockutils [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1181.377059] env[61907]: DEBUG nova.compute.manager [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] No waiting events found dispatching network-vif-plugged-e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1181.377256] env[61907]: WARNING nova.compute.manager [req-f19b1999-2f8c-4591-9176-c581db7b00ca req-597d3b46-a774-4276-a032-7c46db873f20 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Received unexpected event network-vif-plugged-e20f2645-1b97-4510-9141-1a6888ff9312 for instance with vm_state building and task_state spawning. [ 1181.461475] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Successfully updated port: e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1181.840591] env[61907]: DEBUG nova.network.neutron [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1181.867272] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff85f17e-d343-4b66-b281-603c40065fa6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.877264] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18cb604-e497-4665-8bcf-03c293048058 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.900625] env[61907]: DEBUG nova.compute.manager [req-34bf3580-a8ad-45d6-90ba-aa02974ef04c req-205550b4-3864-45c5-ace8-fbe4a14380fb service nova] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Detach interface failed, port_id=e21af03c-7e0e-4275-b1c9-a14a3accb004, reason: Instance ad8a7d32-6ef3-44d2-9617-49335af29ae2 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1181.964055] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1181.964243] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1181.964421] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1182.342947] env[61907]: INFO nova.compute.manager [-] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Took 1.22 seconds to deallocate network for instance. [ 1182.493318] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1182.624836] env[61907]: DEBUG nova.network.neutron [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1182.850182] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1182.850513] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1182.850780] env[61907]: DEBUG nova.objects.instance [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid ad8a7d32-6ef3-44d2-9617-49335af29ae2 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1183.127074] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.127394] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Instance network_info: |[{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1183.127847] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:76:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e20f2645-1b97-4510-9141-1a6888ff9312', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1183.135170] env[61907]: DEBUG oslo.service.loopingcall [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1183.135385] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1183.135618] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1393039-d61b-4b40-a58e-37e6a3390a74 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.155181] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1183.155181] env[61907]: value = "task-1244501" [ 1183.155181] env[61907]: _type = "Task" [ 1183.155181] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.162800] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244501, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.403424] env[61907]: DEBUG nova.compute.manager [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Received event network-changed-e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1183.403635] env[61907]: DEBUG nova.compute.manager [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Refreshing instance network info cache due to event network-changed-e20f2645-1b97-4510-9141-1a6888ff9312. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1183.403885] env[61907]: DEBUG oslo_concurrency.lockutils [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.404111] env[61907]: DEBUG oslo_concurrency.lockutils [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.404419] env[61907]: DEBUG nova.network.neutron [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Refreshing network info cache for port e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1183.434462] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91898a9b-5219-4d3c-a132-84b400290a4c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.442825] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f759b587-d0a3-4529-a830-6b964b347f73 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.472512] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b719e79a-a505-4e44-9fef-8979cc7ee4da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.479735] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59ef6878-2d23-4ac0-a33c-b3ead78ee84a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.492334] env[61907]: DEBUG nova.compute.provider_tree [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1183.665955] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244501, 'name': CreateVM_Task, 'duration_secs': 0.309757} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.666141] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1183.666800] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.666973] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.667305] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1183.667545] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6fbbd2fb-e4b6-4069-bcff-999f5184b2da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.671876] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1183.671876] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5221eee0-05e8-a415-e16c-823a43eb0e99" [ 1183.671876] env[61907]: _type = "Task" [ 1183.671876] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.678866] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5221eee0-05e8-a415-e16c-823a43eb0e99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.994824] env[61907]: DEBUG nova.scheduler.client.report [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1184.112886] env[61907]: DEBUG nova.network.neutron [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updated VIF entry in instance network info cache for port e20f2645-1b97-4510-9141-1a6888ff9312. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1184.113288] env[61907]: DEBUG nova.network.neutron [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.182616] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5221eee0-05e8-a415-e16c-823a43eb0e99, 'name': SearchDatastore_Task, 'duration_secs': 0.012503} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.182915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.183206] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1184.183492] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1184.183673] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1184.183892] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1184.184214] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9fa0dfa-1ba2-4ff6-b8b7-315d9c36d0c2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.191818] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1184.191994] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1184.192670] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7123d6d-65cd-4447-b485-8fd8ce27d54f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.197448] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1184.197448] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e9b86-718e-09bc-e803-4631947816d2" [ 1184.197448] env[61907]: _type = "Task" [ 1184.197448] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.204233] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e9b86-718e-09bc-e803-4631947816d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.500102] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1184.519024] env[61907]: INFO nova.scheduler.client.report [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance ad8a7d32-6ef3-44d2-9617-49335af29ae2 [ 1184.616305] env[61907]: DEBUG oslo_concurrency.lockutils [req-c8509ad2-b145-4e5f-8390-657186bfafd7 req-c5bd88db-a133-4df4-a71a-7a917465ebea service nova] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1184.708202] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529e9b86-718e-09bc-e803-4631947816d2, 'name': SearchDatastore_Task, 'duration_secs': 0.008414} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.708977] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-266772b7-ee4e-4a7d-a81e-c62ec204ac36 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.713759] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1184.713759] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ad0f7-4c3a-249b-105c-5eb7deab9f58" [ 1184.713759] env[61907]: _type = "Task" [ 1184.713759] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.720887] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ad0f7-4c3a-249b-105c-5eb7deab9f58, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.027237] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d055033a-0937-4fc3-9cba-467cd0b34518 tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "ad8a7d32-6ef3-44d2-9617-49335af29ae2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.005s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.224953] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]523ad0f7-4c3a-249b-105c-5eb7deab9f58, 'name': SearchDatastore_Task, 'duration_secs': 0.009538} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.225263] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.226071] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1185.226071] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-abfa3c7c-698d-48d8-97a1-bc26765217d4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.232185] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1185.232185] env[61907]: value = "task-1244502" [ 1185.232185] env[61907]: _type = "Task" [ 1185.232185] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.240075] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.741995] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417749} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.742288] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1185.742505] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1185.742753] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-97ed17da-435f-4ca8-a300-1e7f5a919d76 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.744590] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "b2c5689e-f930-4548-b061-fd2974d44575" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.744820] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.746321] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "b2c5689e-f930-4548-b061-fd2974d44575-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.746528] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.746727] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.748759] env[61907]: INFO nova.compute.manager [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Terminating instance [ 1185.752585] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1185.752585] env[61907]: value = "task-1244503" [ 1185.752585] env[61907]: _type = "Task" [ 1185.752585] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.759812] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244503, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.252974] env[61907]: DEBUG nova.compute.manager [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1186.253292] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.254040] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-effe27d6-c49a-4613-8b73-93274a22dbed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.264837] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244503, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.333178} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.266761] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1186.267111] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1186.267801] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0bba20-91a6-4601-aa3d-aa27beed1d0c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.270018] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6eae9e06-e263-409b-ac4d-0024d80dfd68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.290448] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1186.291719] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-02ccf4aa-2095-4a83-8ac2-70e2c5124446 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.305651] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1186.305651] env[61907]: value = "task-1244504" [ 1186.305651] env[61907]: _type = "Task" [ 1186.305651] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.310396] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1186.310396] env[61907]: value = "task-1244505" [ 1186.310396] env[61907]: _type = "Task" [ 1186.310396] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.316410] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244504, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.321442] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244505, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.818031] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244504, 'name': PowerOffVM_Task, 'duration_secs': 0.220905} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.818031] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.818031] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1186.818692] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fa8b2e76-84e9-4fc5-9b2e-7a33f17525de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.822690] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244505, 'name': ReconfigVM_Task, 'duration_secs': 0.295727} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.823374] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1186.824118] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7708f349-6fc1-4f99-adab-747d1a4b8733 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.830120] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1186.830120] env[61907]: value = "task-1244507" [ 1186.830120] env[61907]: _type = "Task" [ 1186.830120] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.837136] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244507, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.889369] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1186.889642] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1186.889813] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleting the datastore file [datastore1] b2c5689e-f930-4548-b061-fd2974d44575 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1186.890125] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1186541b-c960-4e4b-8dc7-25d7effbe9fb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.896113] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for the task: (returnval){ [ 1186.896113] env[61907]: value = "task-1244508" [ 1186.896113] env[61907]: _type = "Task" [ 1186.896113] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.904115] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.339392] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244507, 'name': Rename_Task, 'duration_secs': 0.160428} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.339748] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1187.339924] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29902671-8604-4071-9c8e-d3056972290e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1187.345612] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1187.345612] env[61907]: value = "task-1244509" [ 1187.345612] env[61907]: _type = "Task" [ 1187.345612] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1187.352940] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.405867] env[61907]: DEBUG oslo_vmware.api [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Task: {'id': task-1244508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132898} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.406166] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1187.406365] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1187.406547] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1187.406740] env[61907]: INFO nova.compute.manager [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1187.407017] env[61907]: DEBUG oslo.service.loopingcall [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1187.407235] env[61907]: DEBUG nova.compute.manager [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1187.407336] env[61907]: DEBUG nova.network.neutron [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1187.739956] env[61907]: DEBUG nova.compute.manager [req-eebce112-9556-49d1-a96d-376f19c3d88b req-87f6a6ee-3ff8-46a9-a6b8-f61b5219afb3 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Received event network-vif-deleted-b7665409-68a9-43d3-a59f-9a2562dc8923 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1187.740186] env[61907]: INFO nova.compute.manager [req-eebce112-9556-49d1-a96d-376f19c3d88b req-87f6a6ee-3ff8-46a9-a6b8-f61b5219afb3 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Neutron deleted interface b7665409-68a9-43d3-a59f-9a2562dc8923; detaching it from the instance and deleting it from the info cache [ 1187.740365] env[61907]: DEBUG nova.network.neutron [req-eebce112-9556-49d1-a96d-376f19c3d88b req-87f6a6ee-3ff8-46a9-a6b8-f61b5219afb3 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1187.855416] env[61907]: DEBUG oslo_vmware.api [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244509, 'name': PowerOnVM_Task, 'duration_secs': 0.442244} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.855684] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1187.855892] env[61907]: INFO nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1187.856100] env[61907]: DEBUG nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1187.856888] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c18773-0c96-47ce-b1b1-efadcfd12f12 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.218015] env[61907]: DEBUG nova.network.neutron [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.244928] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-13110489-d7a1-429a-93c9-c141a8272eef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.254185] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d575c9f-658a-403f-9d5f-06683a2b3aa0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.279033] env[61907]: DEBUG nova.compute.manager [req-eebce112-9556-49d1-a96d-376f19c3d88b req-87f6a6ee-3ff8-46a9-a6b8-f61b5219afb3 service nova] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Detach interface failed, port_id=b7665409-68a9-43d3-a59f-9a2562dc8923, reason: Instance b2c5689e-f930-4548-b061-fd2974d44575 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1188.374943] env[61907]: INFO nova.compute.manager [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Took 11.40 seconds to build instance. [ 1188.720550] env[61907]: INFO nova.compute.manager [-] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Took 1.31 seconds to deallocate network for instance. [ 1188.876907] env[61907]: DEBUG oslo_concurrency.lockutils [None req-3d694ac9-bce7-4529-8e25-49f86df85c3d tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.912s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.209949] env[61907]: DEBUG nova.compute.manager [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Received event network-changed-e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1189.210009] env[61907]: DEBUG nova.compute.manager [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Refreshing instance network info cache due to event network-changed-e20f2645-1b97-4510-9141-1a6888ff9312. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1189.210231] env[61907]: DEBUG oslo_concurrency.lockutils [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1189.210383] env[61907]: DEBUG oslo_concurrency.lockutils [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1189.210550] env[61907]: DEBUG nova.network.neutron [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Refreshing network info cache for port e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1189.226765] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.227043] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.227272] env[61907]: DEBUG nova.objects.instance [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lazy-loading 'resources' on Instance uuid b2c5689e-f930-4548-b061-fd2974d44575 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.829809] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c8045-dad1-46d9-bd6c-a273961809e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.838603] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69fef2f0-5791-43c2-b563-d699fda1a5d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.870131] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b71ef58-025e-4021-9427-771cd3fcaf03 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.877134] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5c6f4ea-ce6d-4dea-8926-677192a9e49e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.890060] env[61907]: DEBUG nova.compute.provider_tree [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1189.932327] env[61907]: DEBUG nova.network.neutron [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updated VIF entry in instance network info cache for port e20f2645-1b97-4510-9141-1a6888ff9312. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1189.932688] env[61907]: DEBUG nova.network.neutron [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1190.393280] env[61907]: DEBUG nova.scheduler.client.report [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1190.435616] env[61907]: DEBUG oslo_concurrency.lockutils [req-0637cb88-755d-4999-9c48-2f335c7819e2 req-0c9aa2af-b103-470a-b5e9-ee44843dcac5 service nova] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1190.898566] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.671s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1190.916185] env[61907]: INFO nova.scheduler.client.report [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Deleted allocations for instance b2c5689e-f930-4548-b061-fd2974d44575 [ 1191.424681] env[61907]: DEBUG oslo_concurrency.lockutils [None req-869f033a-347d-44b7-9cd7-e847fa992b8b tempest-ServersTestJSON-1289139975 tempest-ServersTestJSON-1289139975-project-member] Lock "b2c5689e-f930-4548-b061-fd2974d44575" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.680s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.890103] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.890443] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.890589] env[61907]: DEBUG nova.compute.manager [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1193.891548] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c0351f-3905-41dd-af6b-7ef254aa8466 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1193.898671] env[61907]: DEBUG nova.compute.manager [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61907) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3403}} [ 1193.899336] env[61907]: DEBUG nova.objects.instance [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'flavor' on Instance uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1194.907012] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1194.907361] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10891640-f218-4f3b-a336-9c992f90cc0a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.915309] env[61907]: DEBUG oslo_vmware.api [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1194.915309] env[61907]: value = "task-1244510" [ 1194.915309] env[61907]: _type = "Task" [ 1194.915309] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1194.923045] env[61907]: DEBUG oslo_vmware.api [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244510, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1195.426846] env[61907]: DEBUG oslo_vmware.api [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244510, 'name': PowerOffVM_Task, 'duration_secs': 0.185588} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1195.427204] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1195.427459] env[61907]: DEBUG nova.compute.manager [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1195.428337] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb095b34-33b9-4bcb-aeb9-e669521c444b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1195.939564] env[61907]: DEBUG oslo_concurrency.lockutils [None req-cf01f617-837a-452b-b807-83706ac07149 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.049s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.370338] env[61907]: DEBUG nova.objects.instance [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'flavor' on Instance uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1196.875667] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.875864] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.876020] env[61907]: DEBUG nova.network.neutron [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1196.876212] env[61907]: DEBUG nova.objects.instance [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'info_cache' on Instance uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1197.379930] env[61907]: DEBUG nova.objects.base [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1198.088193] env[61907]: DEBUG nova.network.neutron [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1198.591025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1199.596492] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1199.597016] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8312456-c196-47fd-9056-af7d42cdada5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.604676] env[61907]: DEBUG oslo_vmware.api [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1199.604676] env[61907]: value = "task-1244511" [ 1199.604676] env[61907]: _type = "Task" [ 1199.604676] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1199.612111] env[61907]: DEBUG oslo_vmware.api [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244511, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1200.114988] env[61907]: DEBUG oslo_vmware.api [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244511, 'name': PowerOnVM_Task, 'duration_secs': 0.342972} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1200.115290] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1200.115497] env[61907]: DEBUG nova.compute.manager [None req-0322bf64-8505-4e2f-912b-c311920ff52d tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1200.116273] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5423fb4-6f54-4804-866c-2fbf5a9ce121 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.350263] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea8c2083-44d7-4a0a-af6e-421594aeb6c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.356921] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Suspending the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1201.357175] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-8ebf4dfd-f096-414f-960c-c97d3671dd7b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1201.363510] env[61907]: DEBUG oslo_vmware.api [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1201.363510] env[61907]: value = "task-1244512" [ 1201.363510] env[61907]: _type = "Task" [ 1201.363510] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1201.371319] env[61907]: DEBUG oslo_vmware.api [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244512, 'name': SuspendVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1201.873411] env[61907]: DEBUG oslo_vmware.api [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244512, 'name': SuspendVM_Task} progress is 75%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1202.374048] env[61907]: DEBUG oslo_vmware.api [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244512, 'name': SuspendVM_Task, 'duration_secs': 0.581049} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1202.374491] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Suspended the VM {{(pid=61907) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1202.374491] env[61907]: DEBUG nova.compute.manager [None req-bdc09751-71a0-470f-833f-491b6c542059 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1202.375267] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4a568ee-c4a9-42b5-80e0-42637da91f45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.716234] env[61907]: INFO nova.compute.manager [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Resuming [ 1203.716944] env[61907]: DEBUG nova.objects.instance [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'flavor' on Instance uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1204.726316] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1204.726591] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquired lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1204.726692] env[61907]: DEBUG nova.network.neutron [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1205.450379] env[61907]: DEBUG nova.network.neutron [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [{"id": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "address": "fa:16:3e:e1:3f:fb", "network": {"id": "e0c2d886-5eb9-4d09-8a4e-c437f9e247c8", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-1244255521-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.220", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7652e98cde994af28b7bac0b81547474", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e2153f70-3d14-42ab-8bb3-be78296dd3b8", "external-id": "nsx-vlan-transportzone-532", "segmentation_id": 532, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1a805adb-8e", "ovs_interfaceid": "1a805adb-8e64-4323-8acf-fa4e9fcf7fdf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.953020] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Releasing lock "refresh_cache-f1dc869a-473d-4d7e-bf4b-c46498c55c56" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1205.954096] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-508cf0ab-3885-4743-8ed7-b2c1ca7606a5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.960953] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Resuming the VM {{(pid=61907) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1205.961219] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9a9d5b92-1c4e-46a2-bb39-10ee49bfef0c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.967253] env[61907]: DEBUG oslo_vmware.api [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1205.967253] env[61907]: value = "task-1244513" [ 1205.967253] env[61907]: _type = "Task" [ 1205.967253] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.974890] env[61907]: DEBUG oslo_vmware.api [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.477493] env[61907]: DEBUG oslo_vmware.api [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244513, 'name': PowerOnVM_Task, 'duration_secs': 0.500845} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.477796] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Resumed the VM {{(pid=61907) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1206.478080] env[61907]: DEBUG nova.compute.manager [None req-a65f8f7c-87df-422d-927e-e44d7f94524b tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1206.478829] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2ae74d-a66c-44dd-9a25-bbb8718b4721 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.335476] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.335864] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.335973] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1207.337504] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1207.337701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1207.339759] env[61907]: INFO nova.compute.manager [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Terminating instance [ 1207.843939] env[61907]: DEBUG nova.compute.manager [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1207.844289] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1207.845257] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e1c12a-c8b1-46c1-a49d-c1dc69ec8292 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.853294] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1207.853520] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a3bb1b4-e93a-4a1b-9a66-b537b7dc855e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.858979] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1207.858979] env[61907]: value = "task-1244514" [ 1207.858979] env[61907]: _type = "Task" [ 1207.858979] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.866551] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244514, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.368361] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244514, 'name': PowerOffVM_Task, 'duration_secs': 0.202609} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.368699] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1208.368750] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1208.368985] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de16706b-2f4c-428e-a9e9-8f4b36612917 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.426189] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1208.426445] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1208.426709] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleting the datastore file [datastore2] f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1208.427143] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e7f8c35-8d69-41ae-8565-80140cea4112 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.434210] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for the task: (returnval){ [ 1208.434210] env[61907]: value = "task-1244516" [ 1208.434210] env[61907]: _type = "Task" [ 1208.434210] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.441535] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244516, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.944328] env[61907]: DEBUG oslo_vmware.api [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Task: {'id': task-1244516, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148696} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.944613] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1208.944930] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1208.945268] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1208.945580] env[61907]: INFO nova.compute.manager [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1208.945938] env[61907]: DEBUG oslo.service.loopingcall [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1208.946185] env[61907]: DEBUG nova.compute.manager [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1208.946348] env[61907]: DEBUG nova.network.neutron [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1209.410386] env[61907]: DEBUG nova.compute.manager [req-1e1db2ad-d6cf-44ae-829e-0e54e1b2904a req-a7a28c4c-c871-41cb-a5fc-d01591a893f5 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Received event network-vif-deleted-1a805adb-8e64-4323-8acf-fa4e9fcf7fdf {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1209.411484] env[61907]: INFO nova.compute.manager [req-1e1db2ad-d6cf-44ae-829e-0e54e1b2904a req-a7a28c4c-c871-41cb-a5fc-d01591a893f5 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Neutron deleted interface 1a805adb-8e64-4323-8acf-fa4e9fcf7fdf; detaching it from the instance and deleting it from the info cache [ 1209.411484] env[61907]: DEBUG nova.network.neutron [req-1e1db2ad-d6cf-44ae-829e-0e54e1b2904a req-a7a28c4c-c871-41cb-a5fc-d01591a893f5 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.885030] env[61907]: DEBUG nova.network.neutron [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1209.914175] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c343c01c-d887-4a65-9574-60ebb88b244a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.924653] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a153703c-cdc2-45e5-9e2a-212c91e4cb6d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.948111] env[61907]: DEBUG nova.compute.manager [req-1e1db2ad-d6cf-44ae-829e-0e54e1b2904a req-a7a28c4c-c871-41cb-a5fc-d01591a893f5 service nova] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Detach interface failed, port_id=1a805adb-8e64-4323-8acf-fa4e9fcf7fdf, reason: Instance f1dc869a-473d-4d7e-bf4b-c46498c55c56 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1210.388198] env[61907]: INFO nova.compute.manager [-] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Took 1.44 seconds to deallocate network for instance. [ 1210.894395] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1210.894792] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1210.894919] env[61907]: DEBUG nova.objects.instance [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lazy-loading 'resources' on Instance uuid f1dc869a-473d-4d7e-bf4b-c46498c55c56 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1211.449094] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ff2d59-b798-49d9-8b65-a873a849409e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.456644] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2f63a8-d697-492f-9d2d-41e3acf1f948 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.486252] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496b2136-7e1d-40a1-b180-8ff6c88e916e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.492874] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb5e00b-7de4-41b0-963e-8cda484661db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.505482] env[61907]: DEBUG nova.compute.provider_tree [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1212.008946] env[61907]: DEBUG nova.scheduler.client.report [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1212.514020] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1212.535539] env[61907]: INFO nova.scheduler.client.report [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Deleted allocations for instance f1dc869a-473d-4d7e-bf4b-c46498c55c56 [ 1213.044433] env[61907]: DEBUG oslo_concurrency.lockutils [None req-710ef0d4-1b02-4202-baef-e125ce3e11b0 tempest-ServerActionsTestJSON-1406539900 tempest-ServerActionsTestJSON-1406539900-project-member] Lock "f1dc869a-473d-4d7e-bf4b-c46498c55c56" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.709s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1215.508772] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1217.513834] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.513334] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1218.513566] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1220.021846] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Didn't find any instances for network info cache update. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 1220.022139] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.514016] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1220.514302] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1221.017151] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1221.017401] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1221.017629] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1221.017797] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1221.018741] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc934a8-caf0-4165-b6a0-d089b5ca805e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.027385] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560a6d4b-0a53-48cd-9add-72301945f6b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.041135] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50e79e98-ccbe-4e10-a665-c26b603ed7e0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.047098] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8f260f-7477-4091-aba0-7e542f46de23 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.074543] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180648MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1221.074683] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1221.074869] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1222.102949] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1222.103234] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 97e5b53f-557e-441a-9097-ffdb5a86ad38 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1222.103344] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1222.103491] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1222.139228] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de332a0-9e02-4fbf-827f-46f56b499e99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.146521] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f914f-85c8-42c2-a929-7934bbbf4368 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.175152] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd64a57e-c3c6-488f-ab70-62d6ecae7492 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.181504] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78501f0c-107a-4cb8-a1dc-0d5044355692 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.193754] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1222.696678] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1223.202020] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1223.202445] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.127s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1224.196628] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.702066] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.702455] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.702455] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1224.702455] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1227.542239] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1227.542541] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1228.047051] env[61907]: DEBUG nova.compute.utils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1228.548699] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.607780] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1229.608302] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1229.608392] env[61907]: INFO nova.compute.manager [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Attaching volume 90dd9ce1-874c-4949-9a44-6eb69c30f658 to /dev/sdb [ 1229.638147] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7890996-09a0-4cce-bb08-a94a019c5c3b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.645204] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e8f60e-a599-4f4d-9930-25f66571dfcd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.658515] env[61907]: DEBUG nova.virt.block_device [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating existing volume attachment record: f8a21ee7-2698-4924-a28d-e12f17882a5a {{(pid=61907) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1230.443936] env[61907]: INFO nova.compute.manager [None req-afbd9a52-df45-42b5-b287-c3a19fd4e89e tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Get console output [ 1230.444347] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-afbd9a52-df45-42b5-b287-c3a19fd4e89e tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] The console log is missing. Check your VSPC configuration [ 1231.510650] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.511033] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.511188] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1231.511451] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1231.511640] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.514919] env[61907]: INFO nova.compute.manager [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Terminating instance [ 1232.019326] env[61907]: DEBUG nova.compute.manager [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1232.019702] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1232.020998] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d1dcae-accf-4218-abb8-2217f9a65d8a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.032393] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1232.032706] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b23ddb5-d05e-4b77-aad6-879c0a904b81 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.039246] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1232.039246] env[61907]: value = "task-1244521" [ 1232.039246] env[61907]: _type = "Task" [ 1232.039246] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.049014] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244521, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1232.549210] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244521, 'name': PowerOffVM_Task, 'duration_secs': 0.197764} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1232.549579] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1232.549643] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1232.549899] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d08de1d-7b3f-4868-9ab7-547af39ee577 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.613777] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1232.614040] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1232.614205] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleting the datastore file [datastore2] f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1232.614476] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31fbc9da-8fe0-4cc9-b798-2d5cf3d38f56 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.620416] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1232.620416] env[61907]: value = "task-1244523" [ 1232.620416] env[61907]: _type = "Task" [ 1232.620416] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1232.627507] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244523, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1233.131518] env[61907]: DEBUG oslo_vmware.api [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244523, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123639} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1233.131772] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1233.131966] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1233.132184] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1233.132369] env[61907]: INFO nova.compute.manager [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1233.132615] env[61907]: DEBUG oslo.service.loopingcall [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1233.132819] env[61907]: DEBUG nova.compute.manager [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1233.132914] env[61907]: DEBUG nova.network.neutron [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1233.612419] env[61907]: DEBUG nova.compute.manager [req-be33aa3b-a3af-4875-9103-b9f94e4918f2 req-a9911acb-e30e-415d-8375-7244f515f149 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Received event network-vif-deleted-ed422e68-2fc3-4015-8d2a-716f55d3ea41 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1233.612671] env[61907]: INFO nova.compute.manager [req-be33aa3b-a3af-4875-9103-b9f94e4918f2 req-a9911acb-e30e-415d-8375-7244f515f149 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Neutron deleted interface ed422e68-2fc3-4015-8d2a-716f55d3ea41; detaching it from the instance and deleting it from the info cache [ 1233.612793] env[61907]: DEBUG nova.network.neutron [req-be33aa3b-a3af-4875-9103-b9f94e4918f2 req-a9911acb-e30e-415d-8375-7244f515f149 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.088610] env[61907]: DEBUG nova.network.neutron [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1234.115372] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ada0ffa-fd60-48f3-b0e5-e0cae82d8e96 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.125933] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a52fce60-20c9-4500-b89a-1ce8f85b2a1e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.150059] env[61907]: DEBUG nova.compute.manager [req-be33aa3b-a3af-4875-9103-b9f94e4918f2 req-a9911acb-e30e-415d-8375-7244f515f149 service nova] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Detach interface failed, port_id=ed422e68-2fc3-4015-8d2a-716f55d3ea41, reason: Instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1234.201685] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Volume attach. Driver type: vmdk {{(pid=61907) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1234.201936] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268346', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'name': 'volume-90dd9ce1-874c-4949-9a44-6eb69c30f658', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97e5b53f-557e-441a-9097-ffdb5a86ad38', 'attached_at': '', 'detached_at': '', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'serial': '90dd9ce1-874c-4949-9a44-6eb69c30f658'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1234.202836] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89bf5c3-0030-41e8-a6b0-b2c959cfda0d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.218900] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43491cb-048c-4db2-a87d-7ce5a806fc27 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.244797] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] volume-90dd9ce1-874c-4949-9a44-6eb69c30f658/volume-90dd9ce1-874c-4949-9a44-6eb69c30f658.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1234.245630] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61fb6637-ab29-42bb-bd79-bfe5a720d323 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.263200] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1234.263200] env[61907]: value = "task-1244524" [ 1234.263200] env[61907]: _type = "Task" [ 1234.263200] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.270989] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244524, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1234.591653] env[61907]: INFO nova.compute.manager [-] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Took 1.46 seconds to deallocate network for instance. [ 1234.772642] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244524, 'name': ReconfigVM_Task, 'duration_secs': 0.366381} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1234.772944] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to attach disk [datastore1] volume-90dd9ce1-874c-4949-9a44-6eb69c30f658/volume-90dd9ce1-874c-4949-9a44-6eb69c30f658.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1234.777511] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92c5ef76-172d-462a-a93e-349c0c399bdb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1234.791142] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1234.791142] env[61907]: value = "task-1244525" [ 1234.791142] env[61907]: _type = "Task" [ 1234.791142] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1234.798526] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244525, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1235.098817] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1235.099064] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1235.099306] env[61907]: DEBUG nova.objects.instance [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lazy-loading 'resources' on Instance uuid f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1235.301291] env[61907]: DEBUG oslo_vmware.api [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244525, 'name': ReconfigVM_Task, 'duration_secs': 0.120053} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1235.301601] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268346', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'name': 'volume-90dd9ce1-874c-4949-9a44-6eb69c30f658', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '97e5b53f-557e-441a-9097-ffdb5a86ad38', 'attached_at': '', 'detached_at': '', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'serial': '90dd9ce1-874c-4949-9a44-6eb69c30f658'} {{(pid=61907) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1235.644609] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb32dbd-9cc1-4d3b-9abe-2d86890f1e7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.651969] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7babab30-717a-41c8-9273-f9f8a3205aec {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.680901] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f9fd175-433e-44de-a2d4-f047c30150df {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.687431] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4315ce7f-48c4-45bd-9f73-d91c497cb3b6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.701164] env[61907]: DEBUG nova.compute.provider_tree [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1236.204163] env[61907]: DEBUG nova.scheduler.client.report [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1236.336241] env[61907]: DEBUG nova.objects.instance [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'flavor' on Instance uuid 97e5b53f-557e-441a-9097-ffdb5a86ad38 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1236.708798] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1236.726043] env[61907]: INFO nova.scheduler.client.report [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted allocations for instance f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa [ 1236.842413] env[61907]: DEBUG oslo_concurrency.lockutils [None req-ad54c61a-6284-4676-9b23-d6ecfc141ed0 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.234s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.233499] env[61907]: DEBUG oslo_concurrency.lockutils [None req-d6be783f-a433-4258-8bb6-075893c3a796 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.722s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.764110] env[61907]: DEBUG nova.compute.manager [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Stashing vm_state: active {{(pid=61907) _prep_resize /opt/stack/nova/nova/compute/manager.py:5967}} [ 1238.287379] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1238.287708] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1238.792881] env[61907]: INFO nova.compute.claims [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1239.299296] env[61907]: INFO nova.compute.resource_tracker [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating resource usage from migration f74218e6-c672-4f81-bb59-df3cad250dd8 [ 1239.344212] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b744ad35-4a3d-4f0c-bb38-c70a85ead3da {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.351769] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6490d67c-9f45-4b8b-a99b-43d55309cc6a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.381371] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd275e4d-2d52-4bf9-9bc4-c174ef3a0d6a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.387765] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812cc090-4f53-435d-8b1a-70fa42aa34b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.400183] env[61907]: DEBUG nova.compute.provider_tree [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1239.521200] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "c310aa25-4daa-474b-a801-295269d1b5f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1239.521457] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1239.902952] env[61907]: DEBUG nova.scheduler.client.report [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1240.023476] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1240.407959] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.120s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.408382] env[61907]: INFO nova.compute.manager [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Migrating [ 1240.545185] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.545470] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.546973] env[61907]: INFO nova.compute.claims [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1240.922680] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1240.922982] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1240.923278] env[61907]: DEBUG nova.network.neutron [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1241.604686] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67182827-5b92-4cd7-b624-7324afa66ba9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.612487] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e6a142-accb-4e96-8868-df1e3fe1420c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.642602] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775f20bf-bb46-4bed-a984-571f7a084285 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.649221] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4df011-77e2-4400-b6b8-136298aea235 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.661607] env[61907]: DEBUG nova.compute.provider_tree [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1241.673785] env[61907]: DEBUG nova.network.neutron [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.165081] env[61907]: DEBUG nova.scheduler.client.report [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1242.176056] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.669882] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.124s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1242.670486] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1243.175984] env[61907]: DEBUG nova.compute.utils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1243.177463] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1243.178141] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1243.214599] env[61907]: DEBUG nova.policy [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b4762c5f79e48fb8a4f9701e5c59ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c30836c7053443aaab6e1dbf7783259', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1243.497701] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Successfully created port: 79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1243.681708] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1243.688133] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c3dcfa-21c4-4015-856a-560079fc7a97 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.713025] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 0 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1244.219063] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1244.219845] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-145d9f11-74e9-4073-9473-da4c2cd7f02d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.227071] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1244.227071] env[61907]: value = "task-1244526" [ 1244.227071] env[61907]: _type = "Task" [ 1244.227071] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.234974] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244526, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.690778] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1244.715749] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1244.716031] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1244.716205] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1244.716399] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1244.716553] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1244.718137] env[61907]: DEBUG nova.virt.hardware [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1244.718618] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4256f34-7c63-4bae-8c09-256944649062 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.726173] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45e98e7f-53bb-48c6-9aee-5001b792b1ca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.737573] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244526, 'name': PowerOffVM_Task, 'duration_secs': 0.208287} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.746372] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1244.746578] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 17 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1244.871368] env[61907]: DEBUG nova.compute.manager [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Received event network-vif-plugged-79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1244.871604] env[61907]: DEBUG oslo_concurrency.lockutils [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] Acquiring lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.871825] env[61907]: DEBUG oslo_concurrency.lockutils [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] Lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.872008] env[61907]: DEBUG oslo_concurrency.lockutils [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] Lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.872239] env[61907]: DEBUG nova.compute.manager [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] No waiting events found dispatching network-vif-plugged-79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1244.872396] env[61907]: WARNING nova.compute.manager [req-89f07449-a322-44d4-8336-f43b76ed6a14 req-47eb68f5-a4e3-4d13-83d0-52dd60da6e4e service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Received unexpected event network-vif-plugged-79dafada-2206-4800-8dd1-07aca1f973c4 for instance with vm_state building and task_state spawning. [ 1245.123575] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Successfully updated port: 79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1245.253149] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1245.253401] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1245.253614] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1245.253882] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1245.254065] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1245.254233] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1245.254455] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1245.254624] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1245.255098] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1245.255098] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1245.255262] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1245.260302] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57c4210f-3cb2-49ca-b3a3-cea9f64289de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.280415] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1245.280415] env[61907]: value = "task-1244527" [ 1245.280415] env[61907]: _type = "Task" [ 1245.280415] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1245.293623] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244527, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1245.625847] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1245.626041] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1245.626108] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1245.790144] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244527, 'name': ReconfigVM_Task, 'duration_secs': 0.211117} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1245.790515] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 33 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1246.156779] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1246.275679] env[61907]: DEBUG nova.network.neutron [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1246.296919] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1246.297154] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1246.297369] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1246.297580] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1246.297737] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1246.297890] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1246.298112] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1246.298308] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1246.298502] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1246.298679] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1246.298861] env[61907]: DEBUG nova.virt.hardware [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1246.304186] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1246.304666] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7dc1a0b-a1e0-4b10-bb63-ce3ed82c8f43 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.322921] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1246.322921] env[61907]: value = "task-1244528" [ 1246.322921] env[61907]: _type = "Task" [ 1246.322921] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.330468] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244528, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.778110] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1246.778496] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Instance network_info: |[{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1246.778951] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6a:5d:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79dafada-2206-4800-8dd1-07aca1f973c4', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1246.786296] env[61907]: DEBUG oslo.service.loopingcall [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1246.786512] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1246.786747] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c1a17f92-b099-449c-912e-7b646407f53a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.805951] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1246.805951] env[61907]: value = "task-1244529" [ 1246.805951] env[61907]: _type = "Task" [ 1246.805951] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.813108] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244529, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.830127] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244528, 'name': ReconfigVM_Task, 'duration_secs': 0.181706} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1246.830375] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1246.831086] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ec3552b-b2ba-4783-ae2f-7eb7fcb12178 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.854164] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1246.854426] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d05d285f-950b-444c-b7c9-530755b52f7c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.873049] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1246.873049] env[61907]: value = "task-1244530" [ 1246.873049] env[61907]: _type = "Task" [ 1246.873049] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.881058] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244530, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1246.899368] env[61907]: DEBUG nova.compute.manager [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Received event network-changed-79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1246.899538] env[61907]: DEBUG nova.compute.manager [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Refreshing instance network info cache due to event network-changed-79dafada-2206-4800-8dd1-07aca1f973c4. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1246.899823] env[61907]: DEBUG oslo_concurrency.lockutils [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1246.899990] env[61907]: DEBUG oslo_concurrency.lockutils [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1246.900244] env[61907]: DEBUG nova.network.neutron [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Refreshing network info cache for port 79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1247.317056] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244529, 'name': CreateVM_Task, 'duration_secs': 0.297077} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.317056] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1247.317642] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1247.317831] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1247.318194] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1247.318460] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a8687df7-f0bd-499b-b9e8-01f04adef128 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.322571] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1247.322571] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5265ecfb-54b5-7e8d-6a6b-a2f1a0075172" [ 1247.322571] env[61907]: _type = "Task" [ 1247.322571] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.329471] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5265ecfb-54b5-7e8d-6a6b-a2f1a0075172, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.381239] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244530, 'name': ReconfigVM_Task, 'duration_secs': 0.282611} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.381500] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1247.381791] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 50 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1247.583974] env[61907]: DEBUG nova.network.neutron [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updated VIF entry in instance network info cache for port 79dafada-2206-4800-8dd1-07aca1f973c4. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1247.583974] env[61907]: DEBUG nova.network.neutron [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1247.832914] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5265ecfb-54b5-7e8d-6a6b-a2f1a0075172, 'name': SearchDatastore_Task, 'duration_secs': 0.010767} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.833226] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1247.833446] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1247.833690] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1247.833845] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1247.834046] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1247.834309] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1de42e27-d8d1-43d3-8966-004d2ebb6a8f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.842089] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1247.842271] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1247.842930] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-730645a0-ad10-4086-8433-4f5d3b91518c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.847715] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1247.847715] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52092933-9eeb-ca75-3f19-65f93a583cb4" [ 1247.847715] env[61907]: _type = "Task" [ 1247.847715] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.854598] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52092933-9eeb-ca75-3f19-65f93a583cb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.887810] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f503807-edc1-4ffd-98a3-09764afaae28 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.907716] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05b0c1fa-4312-4a7b-851e-2caaeaf36e40 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.927661] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 67 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1248.086612] env[61907]: DEBUG oslo_concurrency.lockutils [req-438b17c8-e679-4ec0-b3d7-530af63b790a req-2205d9b0-a716-4f03-98ba-3a974ba4a14a service nova] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1248.358261] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52092933-9eeb-ca75-3f19-65f93a583cb4, 'name': SearchDatastore_Task, 'duration_secs': 0.007732} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.358977] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5ea6fbf-4400-4cd0-b020-252899a0822f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.364412] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1248.364412] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b32cd-51c3-1306-7072-a264bc70b348" [ 1248.364412] env[61907]: _type = "Task" [ 1248.364412] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.371654] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b32cd-51c3-1306-7072-a264bc70b348, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1248.875021] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]522b32cd-51c3-1306-7072-a264bc70b348, 'name': SearchDatastore_Task, 'duration_secs': 0.010913} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1248.875309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1248.875309] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] c310aa25-4daa-474b-a801-295269d1b5f0/c310aa25-4daa-474b-a801-295269d1b5f0.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1248.875580] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb7fc2ae-4162-4973-8509-9b6bc21dd2aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.882143] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1248.882143] env[61907]: value = "task-1244531" [ 1248.882143] env[61907]: _type = "Task" [ 1248.882143] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1248.889254] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.392152] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.424976} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.392425] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore1] c310aa25-4daa-474b-a801-295269d1b5f0/c310aa25-4daa-474b-a801-295269d1b5f0.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1249.392643] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1249.392895] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7dda833-ba61-4f5b-a24c-2839e4306a4b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.398441] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1249.398441] env[61907]: value = "task-1244532" [ 1249.398441] env[61907]: _type = "Task" [ 1249.398441] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.405551] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244532, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1249.577154] env[61907]: DEBUG nova.network.neutron [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Port e20f2645-1b97-4510-9141-1a6888ff9312 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1249.908163] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244532, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057847} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1249.908470] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1249.909264] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b079999-9d82-4cbd-93f2-fb5f98ebd691 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.931846] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] c310aa25-4daa-474b-a801-295269d1b5f0/c310aa25-4daa-474b-a801-295269d1b5f0.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1249.932106] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5656ef83-617b-491c-9e70-e1303e28c329 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.950813] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1249.950813] env[61907]: value = "task-1244533" [ 1249.950813] env[61907]: _type = "Task" [ 1249.950813] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1249.958652] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244533, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.462448] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244533, 'name': ReconfigVM_Task, 'duration_secs': 0.267865} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.462740] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Reconfigured VM instance instance-0000006f to attach disk [datastore1] c310aa25-4daa-474b-a801-295269d1b5f0/c310aa25-4daa-474b-a801-295269d1b5f0.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1250.463340] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f9d9e2dc-6326-4cc3-9f47-34335066b0ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.468942] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1250.468942] env[61907]: value = "task-1244534" [ 1250.468942] env[61907]: _type = "Task" [ 1250.468942] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.476127] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244534, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.599554] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.599793] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.599978] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1250.978862] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244534, 'name': Rename_Task, 'duration_secs': 0.132863} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1250.979185] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1250.979414] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6cbc87c0-1bf9-491c-9f5f-7a2c008051d2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.985116] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1250.985116] env[61907]: value = "task-1244535" [ 1250.985116] env[61907]: _type = "Task" [ 1250.985116] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.992198] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244535, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.494683] env[61907]: DEBUG oslo_vmware.api [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244535, 'name': PowerOnVM_Task, 'duration_secs': 0.437696} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1251.495058] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1251.495331] env[61907]: INFO nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1251.495539] env[61907]: DEBUG nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1251.496306] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdad9c3-c5bf-4913-8340-e630907901e9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.634010] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1251.634252] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1251.634451] env[61907]: DEBUG nova.network.neutron [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1252.014861] env[61907]: INFO nova.compute.manager [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Took 11.49 seconds to build instance. [ 1252.408864] env[61907]: DEBUG nova.network.neutron [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1252.517180] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a1a19709-f9fa-4c1c-b700-20d34bfbf7c8 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.995s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.665180] env[61907]: DEBUG nova.compute.manager [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Received event network-changed-79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1252.665411] env[61907]: DEBUG nova.compute.manager [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Refreshing instance network info cache due to event network-changed-79dafada-2206-4800-8dd1-07aca1f973c4. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1252.665642] env[61907]: DEBUG oslo_concurrency.lockutils [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1252.665793] env[61907]: DEBUG oslo_concurrency.lockutils [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1252.665961] env[61907]: DEBUG nova.network.neutron [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Refreshing network info cache for port 79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1252.912054] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1253.410915] env[61907]: DEBUG nova.network.neutron [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updated VIF entry in instance network info cache for port 79dafada-2206-4800-8dd1-07aca1f973c4. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1253.411317] env[61907]: DEBUG nova.network.neutron [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1253.421335] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b959ab-61fe-42e7-9edd-93fdb92b7019 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.429566] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37b65211-4a19-446c-b85f-957103bcb109 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.914269] env[61907]: DEBUG oslo_concurrency.lockutils [req-23708d1a-1ffc-4f41-8e25-dbba2934938e req-ea5be0c1-6617-4a64-a565-f14ba307f4f8 service nova] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1254.531625] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c6e0d00-3af0-46c7-bec1-d3f205606e7d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.553295] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49209a8-ff89-40d9-834b-c19b47eeb5ee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.560201] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 83 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1255.066536] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1255.066854] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ba813a9f-7f0f-4149-92cd-dfef8d6659a1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.074051] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1255.074051] env[61907]: value = "task-1244536" [ 1255.074051] env[61907]: _type = "Task" [ 1255.074051] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.081626] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.583967] env[61907]: DEBUG oslo_vmware.api [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244536, 'name': PowerOnVM_Task, 'duration_secs': 0.401833} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.584290] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1255.584488] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5508d992-1ec9-4aff-b4ea-367f76b36b01 tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance '97e5b53f-557e-441a-9097-ffdb5a86ad38' progress to 100 {{(pid=61907) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1257.474967] env[61907]: DEBUG nova.network.neutron [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Port e20f2645-1b97-4510-9141-1a6888ff9312 binding to destination host cpu-1 is already ACTIVE {{(pid=61907) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1257.475301] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1257.475461] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1257.475635] env[61907]: DEBUG nova.network.neutron [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1258.243869] env[61907]: DEBUG nova.network.neutron [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1258.746935] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1259.250551] env[61907]: DEBUG nova.compute.manager [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=61907) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:900}} [ 1260.345501] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1260.345854] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1260.848574] env[61907]: DEBUG nova.objects.instance [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'migration_context' on Instance uuid 97e5b53f-557e-441a-9097-ffdb5a86ad38 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1261.405609] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e74d62-5136-4674-94e2-a2d9ecda6fbd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.413105] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f057ace-f184-4b5e-adcc-278ffdad32c4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.443953] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5c9d87-ad6f-4417-9292-8bb1662ba2cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.450960] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-972ad6c8-f600-4bd1-8019-5074547fd99e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.463556] env[61907]: DEBUG nova.compute.provider_tree [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1261.966952] env[61907]: DEBUG nova.scheduler.client.report [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1262.978152] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.632s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.511770] env[61907]: INFO nova.compute.manager [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Swapping old allocation on dict_keys(['91bca385-a423-4ca4-9da0-aeb4615e22d3']) held by migration f74218e6-c672-4f81-bb59-df3cad250dd8 for instance [ 1264.532731] env[61907]: DEBUG nova.scheduler.client.report [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Overwriting current allocation {'allocations': {'91bca385-a423-4ca4-9da0-aeb4615e22d3': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 150}}, 'project_id': 'f851fef407334ba2919877cfd4865435', 'user_id': '8186074291264496984da491880a55b9', 'consumer_generation': 1} on consumer 97e5b53f-557e-441a-9097-ffdb5a86ad38 {{(pid=61907) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2033}} [ 1264.614278] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1264.614484] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1264.614670] env[61907]: DEBUG nova.network.neutron [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1265.319663] env[61907]: DEBUG nova.network.neutron [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [{"id": "e20f2645-1b97-4510-9141-1a6888ff9312", "address": "fa:16:3e:54:76:78", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape20f2645-1b", "ovs_interfaceid": "e20f2645-1b97-4510-9141-1a6888ff9312", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1265.823028] env[61907]: DEBUG oslo_concurrency.lockutils [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-97e5b53f-557e-441a-9097-ffdb5a86ad38" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1265.824099] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6a2977-1e52-43a8-b3c4-496a0925cc0c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.831635] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722b6627-be88-4a33-9d6e-653bf8bdb14e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.919031] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1266.919422] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-deefc7c8-0c44-4ef5-81a7-dfd878b1cbbf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.926575] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1266.926575] env[61907]: value = "task-1244537" [ 1266.926575] env[61907]: _type = "Task" [ 1266.926575] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.934571] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.436234] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244537, 'name': PowerOffVM_Task, 'duration_secs': 0.21413} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.436532] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1267.437215] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1267.437445] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1267.437643] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1267.437841] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1267.437996] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1267.438171] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1267.438380] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1267.438561] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1267.438753] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1267.438927] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1267.439122] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1267.444029] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16dedf09-8b8f-431f-8421-10243476d78f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.458610] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1267.458610] env[61907]: value = "task-1244538" [ 1267.458610] env[61907]: _type = "Task" [ 1267.458610] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.465964] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244538, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.968596] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244538, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.469501] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244538, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1268.970339] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244538, 'name': ReconfigVM_Task, 'duration_secs': 1.177904} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.971329] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e5f1cb-4271-4a1c-8c64-0bdd3f277c66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.992488] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1268.992926] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1268.992926] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1268.993118] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1268.993280] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1268.993433] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1268.993657] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1268.993829] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1268.994013] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1268.994199] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1268.994390] env[61907]: DEBUG nova.virt.hardware [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1268.995271] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd87f693-a9e6-414d-b484-5b06de738ba6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.000932] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1269.000932] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52609aab-6e7d-fe2e-e424-2765f276d2c3" [ 1269.000932] env[61907]: _type = "Task" [ 1269.000932] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.010379] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52609aab-6e7d-fe2e-e424-2765f276d2c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1269.510958] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52609aab-6e7d-fe2e-e424-2765f276d2c3, 'name': SearchDatastore_Task, 'duration_secs': 0.008718} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1269.516106] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1269.516401] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43eac487-dd8c-44cd-a07f-bb07e59df7d8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.534671] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1269.534671] env[61907]: value = "task-1244539" [ 1269.534671] env[61907]: _type = "Task" [ 1269.534671] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.542548] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244539, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.044341] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244539, 'name': ReconfigVM_Task, 'duration_secs': 0.205544} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.044679] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to detach disk 2000 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1270.045404] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9993bb5b-5750-4ee4-8278-658eeaf6edd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.069354] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1270.069611] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87c5a8e3-2291-425a-9537-d8de5ce2166b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.086806] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1270.086806] env[61907]: value = "task-1244540" [ 1270.086806] env[61907]: _type = "Task" [ 1270.086806] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.093881] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244540, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.595761] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244540, 'name': ReconfigVM_Task, 'duration_secs': 0.247733} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.596036] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38/97e5b53f-557e-441a-9097-ffdb5a86ad38.vmdk or device None with type thin {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1270.596905] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0caf9e3-b5b8-47f9-9d80-52eb129e220d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.616964] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc8be0f4-ce0a-460d-827b-c2feb2441363 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.636047] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62eb9d94-6173-447d-a59e-18666b339ee2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.655114] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3a67b6-19d8-4e21-8290-0d3e84bd0456 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.660996] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1270.661238] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5eee4357-3671-43b4-8288-07445727eb45 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.666807] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1270.666807] env[61907]: value = "task-1244541" [ 1270.666807] env[61907]: _type = "Task" [ 1270.666807] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.674820] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1271.176568] env[61907]: DEBUG oslo_vmware.api [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244541, 'name': PowerOnVM_Task, 'duration_secs': 0.369783} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1271.177027] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1272.218098] env[61907]: INFO nova.compute.manager [None req-7aa3e213-9270-4c33-bd62-e17ea92dd03a tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance to original state: 'active' [ 1273.363812] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.364208] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.364404] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.364600] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.364781] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.366978] env[61907]: INFO nova.compute.manager [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Terminating instance [ 1273.871567] env[61907]: DEBUG nova.compute.manager [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1273.871833] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1273.872182] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3d415682-9043-4386-98af-863519e113e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.879742] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1273.879742] env[61907]: value = "task-1244542" [ 1273.879742] env[61907]: _type = "Task" [ 1273.879742] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.889471] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244542, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.390010] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244542, 'name': PowerOffVM_Task, 'duration_secs': 0.18572} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.390380] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1274.390431] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Volume detach. Driver type: vmdk {{(pid=61907) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1274.390625] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268346', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'name': 'volume-90dd9ce1-874c-4949-9a44-6eb69c30f658', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '97e5b53f-557e-441a-9097-ffdb5a86ad38', 'attached_at': '2024-10-10T13:19:18.000000', 'detached_at': '', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'serial': '90dd9ce1-874c-4949-9a44-6eb69c30f658'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1274.391421] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a533e591-ba54-4b0c-aeec-976c88b75363 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.411734] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f80c2d49-5792-4bc1-a6f6-4be95a70c4d1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.419452] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c281bba3-f652-4a44-83ab-d728f8fa1ff9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.438852] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5066b72-f7ab-498c-b570-b2c0d2b81616 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.452525] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] The volume has not been displaced from its original location: [datastore1] volume-90dd9ce1-874c-4949-9a44-6eb69c30f658/volume-90dd9ce1-874c-4949-9a44-6eb69c30f658.vmdk. No consolidation needed. {{(pid=61907) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1274.457711] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1274.457982] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3fb8b9c-d825-4fb6-b27c-e5a4bc1f9e4a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.475206] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1274.475206] env[61907]: value = "task-1244543" [ 1274.475206] env[61907]: _type = "Task" [ 1274.475206] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.482192] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.984206] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244543, 'name': ReconfigVM_Task, 'duration_secs': 0.180713} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.984503] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61907) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1274.989144] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4d993d1-a3bf-4d36-b50a-1d7443dc120d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.002950] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1275.002950] env[61907]: value = "task-1244544" [ 1275.002950] env[61907]: _type = "Task" [ 1275.002950] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.015813] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244544, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.512845] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244544, 'name': ReconfigVM_Task, 'duration_secs': 0.128065} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.513261] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-268346', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'name': 'volume-90dd9ce1-874c-4949-9a44-6eb69c30f658', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '97e5b53f-557e-441a-9097-ffdb5a86ad38', 'attached_at': '2024-10-10T13:19:18.000000', 'detached_at': '', 'volume_id': '90dd9ce1-874c-4949-9a44-6eb69c30f658', 'serial': '90dd9ce1-874c-4949-9a44-6eb69c30f658'} {{(pid=61907) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1275.513465] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1275.514236] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b39060-0ca9-48be-8e3a-64fa83ee5eca {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.520357] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1275.520576] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49344b08-247e-41f4-869b-9029c8fe309f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.582285] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1275.582545] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1275.582735] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleting the datastore file [datastore2] 97e5b53f-557e-441a-9097-ffdb5a86ad38 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1275.583018] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bdb0fd76-f51d-48b3-afd2-88e178b802dd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.588683] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1275.588683] env[61907]: value = "task-1244546" [ 1275.588683] env[61907]: _type = "Task" [ 1275.588683] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.595866] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244546, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.098379] env[61907]: DEBUG oslo_vmware.api [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244546, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.138105} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.098646] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1276.098840] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1276.099031] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1276.099229] env[61907]: INFO nova.compute.manager [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Took 2.23 seconds to destroy the instance on the hypervisor. [ 1276.099477] env[61907]: DEBUG oslo.service.loopingcall [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1276.099674] env[61907]: DEBUG nova.compute.manager [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1276.099771] env[61907]: DEBUG nova.network.neutron [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1276.601023] env[61907]: DEBUG nova.compute.manager [req-7d3f7ebe-7991-47f2-a41e-e54317ffdfaf req-fc6cbf36-0d44-4cc5-8639-a9443d959459 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Received event network-vif-deleted-e20f2645-1b97-4510-9141-1a6888ff9312 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1276.601272] env[61907]: INFO nova.compute.manager [req-7d3f7ebe-7991-47f2-a41e-e54317ffdfaf req-fc6cbf36-0d44-4cc5-8639-a9443d959459 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Neutron deleted interface e20f2645-1b97-4510-9141-1a6888ff9312; detaching it from the instance and deleting it from the info cache [ 1276.601438] env[61907]: DEBUG nova.network.neutron [req-7d3f7ebe-7991-47f2-a41e-e54317ffdfaf req-fc6cbf36-0d44-4cc5-8639-a9443d959459 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1277.076057] env[61907]: DEBUG nova.network.neutron [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1277.104572] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1597f96-411f-40eb-9095-9269e50b7de4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.115535] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc66b71-06a9-46db-b1f1-35e3e68e2383 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1277.137623] env[61907]: DEBUG nova.compute.manager [req-7d3f7ebe-7991-47f2-a41e-e54317ffdfaf req-fc6cbf36-0d44-4cc5-8639-a9443d959459 service nova] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Detach interface failed, port_id=e20f2645-1b97-4510-9141-1a6888ff9312, reason: Instance 97e5b53f-557e-441a-9097-ffdb5a86ad38 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1277.579406] env[61907]: INFO nova.compute.manager [-] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Took 1.48 seconds to deallocate network for instance. [ 1278.014642] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1278.121555] env[61907]: INFO nova.compute.manager [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Took 0.54 seconds to detach 1 volumes for instance. [ 1278.627552] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1278.627897] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1278.628795] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1278.653444] env[61907]: INFO nova.scheduler.client.report [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted allocations for instance 97e5b53f-557e-441a-9097-ffdb5a86ad38 [ 1279.161145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a4d3cb4c-908a-4c7c-8804-caf26d94c3bf tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "97e5b53f-557e-441a-9097-ffdb5a86ad38" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.797s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1279.514082] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1279.514474] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.514585] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1280.514969] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1280.515250] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1280.567426] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1280.567701] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.045879] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1281.046069] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1281.046227] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1281.046387] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid c310aa25-4daa-474b-a801-295269d1b5f0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1281.069530] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1281.590025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1281.590296] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1281.592068] env[61907]: INFO nova.compute.claims [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1282.637612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99254145-bbe9-4c78-9055-8b3175113793 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.647284] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f88e7de-c1a4-4596-9f06-587aff6b6c2e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.677976] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e3561c-1823-4112-b211-bfe374ed27a2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.684985] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93955517-c180-422a-9fa1-e874686bf80d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1282.698046] env[61907]: DEBUG nova.compute.provider_tree [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1282.798255] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1283.201453] env[61907]: DEBUG nova.scheduler.client.report [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1283.300550] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1283.300794] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1283.300964] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.301160] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.301322] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1283.711030] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.711030] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1283.804537] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.805335] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1283.805335] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1283.805488] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1283.806436] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6e338d-7ca6-4598-b72e-fbc6a2f95959 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.814809] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bbb80d-dbfb-432f-86f1-e04ae3d5b9bd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.828047] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33ff38f-8c47-407e-98c5-0372cac23687 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.833979] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243f2ef1-5fc9-4bab-9c0b-0a9803f78e85 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1283.861913] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181260MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1283.862061] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1283.862254] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1284.216730] env[61907]: DEBUG nova.compute.utils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1284.218579] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1284.218758] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1284.256131] env[61907]: DEBUG nova.policy [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8186074291264496984da491880a55b9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f851fef407334ba2919877cfd4865435', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1284.557132] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Successfully created port: 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1284.721765] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1284.886932] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c310aa25-4daa-474b-a801-295269d1b5f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1284.887107] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 26db17bf-ac9c-403f-b4fd-040db331706b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1284.887299] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1284.887441] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1284.921804] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb612ea-0195-4a36-abf7-c6208a1c2065 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.928943] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e0f4a6-1b8c-422c-8d92-92def893a466 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.959041] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eab9ee1-9b3f-4ce8-9f30-21f979810f5e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.965593] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caddbc7a-bd0b-4b3e-ba7e-6644a5e69358 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1284.977975] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1285.481113] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1285.732312] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1285.758569] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1285.758862] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1285.759099] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1285.759309] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1285.759466] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1285.759617] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1285.759830] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1285.760107] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1285.760294] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1285.760466] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1285.760645] env[61907]: DEBUG nova.virt.hardware [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1285.761505] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26257e11-a3c6-4989-9847-19d742ed3b0f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.769216] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58fac03-29fb-4c9f-9a75-d99e64448fc3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1285.937703] env[61907]: DEBUG nova.compute.manager [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1285.937927] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1285.938165] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1285.938343] env[61907]: DEBUG oslo_concurrency.lockutils [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1285.938518] env[61907]: DEBUG nova.compute.manager [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] No waiting events found dispatching network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1285.938692] env[61907]: WARNING nova.compute.manager [req-f4f7986d-e608-4982-ad6a-b9c2d8e32637 req-29298925-d0bd-42b0-ac89-f4e83d0516e4 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received unexpected event network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 for instance with vm_state building and task_state spawning. [ 1285.986247] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1285.986411] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.124s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1286.022865] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Successfully updated port: 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1286.198796] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.199067] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1286.199205] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1286.525338] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1286.525499] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1286.525638] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1287.056964] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1287.179863] env[61907]: DEBUG nova.network.neutron [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1287.682584] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1287.682934] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance network_info: |[{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1287.683434] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:61:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ff33968-4c0e-40ce-97bf-32051f27be74', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1287.690820] env[61907]: DEBUG oslo.service.loopingcall [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1287.691046] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1287.691279] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4a54f897-4670-47b0-9d88-c4bdfbffa0de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1287.711404] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1287.711404] env[61907]: value = "task-1244547" [ 1287.711404] env[61907]: _type = "Task" [ 1287.711404] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1287.718368] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244547, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1287.968974] env[61907]: DEBUG nova.compute.manager [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1287.969161] env[61907]: DEBUG nova.compute.manager [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing instance network info cache due to event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1287.969387] env[61907]: DEBUG oslo_concurrency.lockutils [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1287.969540] env[61907]: DEBUG oslo_concurrency.lockutils [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1287.969703] env[61907]: DEBUG nova.network.neutron [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1288.221891] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244547, 'name': CreateVM_Task, 'duration_secs': 0.278854} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.222271] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1288.222713] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.222893] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.223267] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1288.223524] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33b220fd-8f91-4607-9d02-360d50d4c619 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.227636] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1288.227636] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520a436a-7169-8b25-60bf-232cbdefec90" [ 1288.227636] env[61907]: _type = "Task" [ 1288.227636] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.234900] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520a436a-7169-8b25-60bf-232cbdefec90, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1288.693518] env[61907]: DEBUG nova.network.neutron [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updated VIF entry in instance network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1288.693734] env[61907]: DEBUG nova.network.neutron [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1288.737984] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]520a436a-7169-8b25-60bf-232cbdefec90, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1288.738322] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1288.738558] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1288.738793] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1288.738948] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1288.739152] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1288.739505] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7895ebff-a670-4ab1-9187-dde30c9b16e4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.747432] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1288.747604] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1288.748351] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e389594f-4f45-42f1-bd58-3ab469e5d040 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1288.753029] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1288.753029] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5240122d-5fee-3da4-b0d3-127b8342d36c" [ 1288.753029] env[61907]: _type = "Task" [ 1288.753029] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1288.759677] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5240122d-5fee-3da4-b0d3-127b8342d36c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.196263] env[61907]: DEBUG oslo_concurrency.lockutils [req-6433df11-bf0f-4fc6-9f7a-e8c3166d61cc req-5caf0890-7e27-4a7f-9f95-68e2851c2021 service nova] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.264727] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5240122d-5fee-3da4-b0d3-127b8342d36c, 'name': SearchDatastore_Task, 'duration_secs': 0.007746} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.265512] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6428d5c-0047-45c2-aec8-8736ac24de7e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.270311] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1289.270311] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b923bf-b91a-1dcd-24eb-1b87fa74b5b5" [ 1289.270311] env[61907]: _type = "Task" [ 1289.270311] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.277213] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b923bf-b91a-1dcd-24eb-1b87fa74b5b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1289.781602] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52b923bf-b91a-1dcd-24eb-1b87fa74b5b5, 'name': SearchDatastore_Task, 'duration_secs': 0.008551} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1289.781837] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1289.782115] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1289.782379] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2e3c0858-51a4-4d3a-999e-c17370a4faae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1289.788782] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1289.788782] env[61907]: value = "task-1244548" [ 1289.788782] env[61907]: _type = "Task" [ 1289.788782] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1289.796118] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244548, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.298520] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244548, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.420709} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.298926] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1290.298996] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1290.299321] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab595113-49d8-4187-8fae-14a23f63ecab {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.306378] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1290.306378] env[61907]: value = "task-1244549" [ 1290.306378] env[61907]: _type = "Task" [ 1290.306378] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.314177] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1290.816424] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054992} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1290.816704] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1290.817474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f8da02-5eac-4a6b-9fa7-c590d8fa3064 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.838774] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1290.839043] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85c33145-6b2b-4695-905e-b136340b6da3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1290.859862] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1290.859862] env[61907]: value = "task-1244550" [ 1290.859862] env[61907]: _type = "Task" [ 1290.859862] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1290.867457] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244550, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.369874] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244550, 'name': ReconfigVM_Task, 'duration_secs': 0.267026} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.370282] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1291.370803] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d04df56e-416c-45a5-8e01-5934b915d335 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.376937] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1291.376937] env[61907]: value = "task-1244551" [ 1291.376937] env[61907]: _type = "Task" [ 1291.376937] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.384434] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244551, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1291.887085] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244551, 'name': Rename_Task, 'duration_secs': 0.168562} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1291.887388] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1291.887672] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a489c086-45c0-4d1a-bb52-f2cb5e7bd0f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1291.893038] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1291.893038] env[61907]: value = "task-1244552" [ 1291.893038] env[61907]: _type = "Task" [ 1291.893038] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1291.899897] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244552, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1292.403440] env[61907]: DEBUG oslo_vmware.api [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244552, 'name': PowerOnVM_Task, 'duration_secs': 0.396187} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1292.403789] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1292.403969] env[61907]: INFO nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Took 6.67 seconds to spawn the instance on the hypervisor. [ 1292.404083] env[61907]: DEBUG nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1292.404848] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9f704fc-90f8-4ab7-8a20-ad23b958215b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1292.923920] env[61907]: INFO nova.compute.manager [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Took 11.35 seconds to build instance. [ 1293.426410] env[61907]: DEBUG oslo_concurrency.lockutils [None req-b3e6c770-80a4-4bd8-b794-c84ad04a2a5b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.858s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1293.852546] env[61907]: DEBUG nova.compute.manager [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1293.852755] env[61907]: DEBUG nova.compute.manager [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing instance network info cache due to event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1293.852991] env[61907]: DEBUG oslo_concurrency.lockutils [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1293.853280] env[61907]: DEBUG oslo_concurrency.lockutils [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1293.853462] env[61907]: DEBUG nova.network.neutron [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1294.556162] env[61907]: DEBUG nova.network.neutron [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updated VIF entry in instance network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1294.556567] env[61907]: DEBUG nova.network.neutron [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1295.059602] env[61907]: DEBUG oslo_concurrency.lockutils [req-170c1588-0893-405e-8453-b2cc3aecf005 req-1db090a3-6d65-46ec-a0e2-27fba8ae6515 service nova] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1331.573088] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.573461] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.573547] env[61907]: INFO nova.compute.manager [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Shelving [ 1332.582909] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1332.583242] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f5d9c39-eaeb-4d24-96dd-ab9228a4ff5b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1332.593107] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1332.593107] env[61907]: value = "task-1244553" [ 1332.593107] env[61907]: _type = "Task" [ 1332.593107] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1332.601728] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244553, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1333.102894] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244553, 'name': PowerOffVM_Task, 'duration_secs': 0.178687} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1333.103184] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1333.103946] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a112ac2d-91b4-4ef8-acfa-5dad16ac66fc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.123149] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c59572-8ac7-4e6a-8a05-c45f46641dc1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.634651] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Creating Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1333.635038] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-cd4657d8-edc0-45ff-9b42-69a11699c913 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.642079] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1333.642079] env[61907]: value = "task-1244554" [ 1333.642079] env[61907]: _type = "Task" [ 1333.642079] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1333.650655] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244554, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.152693] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244554, 'name': CreateSnapshot_Task, 'duration_secs': 0.402751} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.152972] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Created Snapshot of the VM instance {{(pid=61907) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1334.153714] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f40655-db17-4c49-8d8f-c246397fd88d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.670985] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Creating linked-clone VM from snapshot {{(pid=61907) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1334.671370] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d45b825f-c691-43e0-b4c8-c4fdc2cb2b48 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.679471] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1334.679471] env[61907]: value = "task-1244555" [ 1334.679471] env[61907]: _type = "Task" [ 1334.679471] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.687392] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244555, 'name': CloneVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.189144] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244555, 'name': CloneVM_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1335.689707] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244555, 'name': CloneVM_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1336.191388] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244555, 'name': CloneVM_Task, 'duration_secs': 1.04911} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1336.191735] env[61907]: INFO nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Created linked-clone VM from snapshot [ 1336.192482] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a08469-c7a8-415a-9340-356a77df16cd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.198966] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Uploading image aefba674-8a3f-4112-8a55-f47a5043daee {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1336.218595] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1336.218595] env[61907]: value = "vm-268350" [ 1336.218595] env[61907]: _type = "VirtualMachine" [ 1336.218595] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1336.218820] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-938eef9e-601b-4d07-8f04-d17792b62fd2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.224676] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease: (returnval){ [ 1336.224676] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ead7b-7729-de1e-24ec-c1c66cc0c82c" [ 1336.224676] env[61907]: _type = "HttpNfcLease" [ 1336.224676] env[61907]: } obtained for exporting VM: (result){ [ 1336.224676] env[61907]: value = "vm-268350" [ 1336.224676] env[61907]: _type = "VirtualMachine" [ 1336.224676] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1336.224920] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the lease: (returnval){ [ 1336.224920] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ead7b-7729-de1e-24ec-c1c66cc0c82c" [ 1336.224920] env[61907]: _type = "HttpNfcLease" [ 1336.224920] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1336.230559] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1336.230559] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ead7b-7729-de1e-24ec-c1c66cc0c82c" [ 1336.230559] env[61907]: _type = "HttpNfcLease" [ 1336.230559] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1336.733095] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1336.733095] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ead7b-7729-de1e-24ec-c1c66cc0c82c" [ 1336.733095] env[61907]: _type = "HttpNfcLease" [ 1336.733095] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1336.733508] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1336.733508] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]528ead7b-7729-de1e-24ec-c1c66cc0c82c" [ 1336.733508] env[61907]: _type = "HttpNfcLease" [ 1336.733508] env[61907]: }. {{(pid=61907) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1336.734116] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4004f2-e954-4b53-adaa-8b34ab6181ad {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1336.741070] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1336.741259] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk for reading. {{(pid=61907) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1336.829028] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b12672b6-e7a8-4645-8105-be5265ed5a9c {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1338.510735] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1340.514424] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1341.508967] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.014323] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1342.014701] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1342.014701] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1342.545035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1342.545035] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1342.546231] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1342.546231] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid c310aa25-4daa-474b-a801-295269d1b5f0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1344.276196] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.515898] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1344.516869] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f6b2ab-309b-49ca-a447-505ebeca57ef {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.522976] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1344.523160] env[61907]: ERROR oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk due to incomplete transfer. [ 1344.523377] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f067c892-ddf6-4d4d-8041-fb23428f381f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.529773] env[61907]: DEBUG oslo_vmware.rw_handles [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5266dcef-4c31-0868-598b-f79f5ec04d32/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1344.529967] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Uploaded image aefba674-8a3f-4112-8a55-f47a5043daee to the Glance image server {{(pid=61907) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1344.532135] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Destroying the VM {{(pid=61907) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1344.532362] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-92a8b3be-567b-4283-9aaf-867c74a156f5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.538828] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1344.538828] env[61907]: value = "task-1244557" [ 1344.538828] env[61907]: _type = "Task" [ 1344.538828] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.546597] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244557, 'name': Destroy_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.779121] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1344.779121] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1344.779121] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.779369] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.779447] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.779624] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1344.779750] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1344.779895] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1345.048814] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244557, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.282972] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.283347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.283402] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1345.283529] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1345.284409] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd84378-7174-4319-80ce-f9cc996104de {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.292573] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6c8a3a-57f0-4ebc-9e30-8e49ae7b5197 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.305879] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9584425-b32c-44cb-b543-7893e00d59b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.311859] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c319ef3-ce40-49c7-b609-08ea99c7ef2a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.340744] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181155MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1345.340893] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1345.341082] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1345.548810] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244557, 'name': Destroy_Task} progress is 100%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.049599] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244557, 'name': Destroy_Task, 'duration_secs': 1.324061} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.049919] env[61907]: INFO nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Destroyed the VM [ 1346.050214] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleting Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1346.050499] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4ee97241-e635-43af-bdca-c133d85825c6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.056407] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1346.056407] env[61907]: value = "task-1244558" [ 1346.056407] env[61907]: _type = "Task" [ 1346.056407] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.063605] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244558, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.367109] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c310aa25-4daa-474b-a801-295269d1b5f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1346.367496] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 26db17bf-ac9c-403f-b4fd-040db331706b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1346.367658] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1346.367869] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1346.383512] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1346.396030] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1346.396030] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1346.405900] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1346.422173] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1346.454068] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b77e68-11a6-4b15-9ad9-3da0e3c6654e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.461536] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caa4d7c-e521-47f1-9caf-16ac422b3903 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.492223] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee572ae2-b833-48ac-9889-76a052a5209a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.498906] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72074ee-c02d-42b4-be51-3413ad079739 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.511612] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1346.564041] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244558, 'name': RemoveSnapshot_Task, 'duration_secs': 0.308952} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.564291] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleted Snapshot of the VM instance {{(pid=61907) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1346.564553] env[61907]: DEBUG nova.compute.manager [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1346.565281] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0e028b-0386-4aca-aeef-bb85e05beea7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.015097] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1347.075954] env[61907]: INFO nova.compute.manager [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Shelve offloading [ 1347.519807] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1347.520213] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.179s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1347.579898] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1347.580232] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-215896c5-2bcb-409f-982e-a09ed05d28ae {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.586886] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1347.586886] env[61907]: value = "task-1244559" [ 1347.586886] env[61907]: _type = "Task" [ 1347.586886] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.596543] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] VM already powered off {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1347.596763] env[61907]: DEBUG nova.compute.manager [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1347.597474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30046cc-51d5-42fa-a006-10c83e069010 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1347.602869] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1347.603057] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1347.603240] env[61907]: DEBUG nova.network.neutron [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1348.254289] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1348.306198] env[61907]: DEBUG nova.network.neutron [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1348.809250] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1349.013462] env[61907]: DEBUG nova.compute.manager [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-vif-unplugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1349.013702] env[61907]: DEBUG oslo_concurrency.lockutils [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.013923] env[61907]: DEBUG oslo_concurrency.lockutils [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.014194] env[61907]: DEBUG oslo_concurrency.lockutils [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1349.014301] env[61907]: DEBUG nova.compute.manager [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] No waiting events found dispatching network-vif-unplugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1349.014462] env[61907]: WARNING nova.compute.manager [req-97b73f47-9453-43e7-a3bc-7f5a46c7aca9 req-8d89477e-194e-4dd8-981d-a211c098655d service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received unexpected event network-vif-unplugged-4ff33968-4c0e-40ce-97bf-32051f27be74 for instance with vm_state shelved and task_state shelving_offloading. [ 1349.158976] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1349.159905] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5489f9ea-3db9-4052-b22d-8443e2194fb9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.167424] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1349.167648] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-485c3f9b-d385-4bee-9991-0ab8fdfecd66 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.242842] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1349.243093] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1349.243294] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleting the datastore file [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1349.243597] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5720995e-1c14-46e2-af41-1d47377d5764 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.250524] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1349.250524] env[61907]: value = "task-1244561" [ 1349.250524] env[61907]: _type = "Task" [ 1349.250524] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.257861] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244561, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.762205] env[61907]: DEBUG oslo_vmware.api [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244561, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151832} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.762476] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1349.762665] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1349.762838] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1349.787838] env[61907]: INFO nova.scheduler.client.report [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted allocations for instance 26db17bf-ac9c-403f-b4fd-040db331706b [ 1350.292357] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1350.292675] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1350.292852] env[61907]: DEBUG nova.objects.instance [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'resources' on Instance uuid 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1350.797029] env[61907]: DEBUG nova.objects.instance [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'numa_topology' on Instance uuid 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1351.041398] env[61907]: DEBUG nova.compute.manager [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1351.041637] env[61907]: DEBUG nova.compute.manager [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing instance network info cache due to event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1351.042206] env[61907]: DEBUG oslo_concurrency.lockutils [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1351.042365] env[61907]: DEBUG oslo_concurrency.lockutils [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1351.042535] env[61907]: DEBUG nova.network.neutron [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1351.299550] env[61907]: DEBUG nova.objects.base [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Object Instance<26db17bf-ac9c-403f-b4fd-040db331706b> lazy-loaded attributes: resources,numa_topology {{(pid=61907) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1351.334306] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932441c8-2b0a-4f32-ae57-8b0ef20b4307 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.342100] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63c4a5f-c1e0-4498-baed-27f7ac1198aa {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.372718] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f257709-6cef-47c4-a789-d93153c78205 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.379400] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e1321b-dc18-4708-8e6c-dbadbbfe86d0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.392124] env[61907]: DEBUG nova.compute.provider_tree [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1351.640894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1351.802271] env[61907]: DEBUG nova.network.neutron [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updated VIF entry in instance network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1351.802650] env[61907]: DEBUG nova.network.neutron [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap4ff33968-4c", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1351.894801] env[61907]: DEBUG nova.scheduler.client.report [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1352.305287] env[61907]: DEBUG oslo_concurrency.lockutils [req-0d3f9894-4c72-4d62-9b0f-6f0fd3e5886a req-891ec43b-1a23-48de-8fde-efd9ef874020 service nova] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1352.399588] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.107s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.907811] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f979dbf1-7faf-4821-85f3-1fedc365639b tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.334s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.908741] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.268s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.908927] env[61907]: INFO nova.compute.manager [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Unshelving [ 1353.932030] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1353.932309] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1353.932537] env[61907]: DEBUG nova.objects.instance [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'pci_requests' on Instance uuid 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1354.438980] env[61907]: DEBUG nova.objects.instance [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'numa_topology' on Instance uuid 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1354.941672] env[61907]: INFO nova.compute.claims [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1355.985627] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e20edef5-1f57-455d-8bc1-30c0b0d61eee {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1355.992939] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d636c6-1e46-43cd-b7fd-3e6c457353e3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.023143] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c73338-5dd2-49e7-b411-2b7c6b61ae2b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.029546] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621315c0-fa1a-4601-9305-13b25ebf78bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.041847] env[61907]: DEBUG nova.compute.provider_tree [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1356.544697] env[61907]: DEBUG nova.scheduler.client.report [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1357.050011] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.117s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.077307] env[61907]: INFO nova.network.neutron [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating port 4ff33968-4c0e-40ce-97bf-32051f27be74 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1358.666399] env[61907]: DEBUG nova.compute.manager [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1358.666626] env[61907]: DEBUG oslo_concurrency.lockutils [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1358.666840] env[61907]: DEBUG oslo_concurrency.lockutils [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1358.667155] env[61907]: DEBUG oslo_concurrency.lockutils [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1358.667414] env[61907]: DEBUG nova.compute.manager [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] No waiting events found dispatching network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1358.667611] env[61907]: WARNING nova.compute.manager [req-d5ae71a2-037e-4cdc-9970-d86e0244381e req-81c7b99d-3db9-44b6-9ed7-eaa352ccf8fe service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received unexpected event network-vif-plugged-4ff33968-4c0e-40ce-97bf-32051f27be74 for instance with vm_state shelved_offloaded and task_state spawning. [ 1358.751290] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1358.751564] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1358.751670] env[61907]: DEBUG nova.network.neutron [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1359.449108] env[61907]: DEBUG nova.network.neutron [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1359.952029] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1359.979132] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='29d0d4390e59732a386a00703f1a5078',container_format='bare',created_at=2024-10-10T13:20:23Z,direct_url=,disk_format='vmdk',id=aefba674-8a3f-4112-8a55-f47a5043daee,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1797779370-shelved',owner='f851fef407334ba2919877cfd4865435',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-10-10T13:20:37Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1359.979410] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1359.979578] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1359.979808] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1359.979975] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1359.980147] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1359.980362] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1359.980528] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1359.980700] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1359.980874] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1359.981066] env[61907]: DEBUG nova.virt.hardware [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1359.982233] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f58e011-1247-49c4-8d27-9be42dce811e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.990046] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5922be4c-ab56-4dd9-ae50-566697bf0719 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.005084] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:61:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ff33968-4c0e-40ce-97bf-32051f27be74', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1360.012358] env[61907]: DEBUG oslo.service.loopingcall [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1360.012579] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1360.012787] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cc6ee61d-fcd9-48dc-97db-4cb58d21effe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.031304] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1360.031304] env[61907]: value = "task-1244562" [ 1360.031304] env[61907]: _type = "Task" [ 1360.031304] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.038021] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244562, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.541449] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244562, 'name': CreateVM_Task, 'duration_secs': 0.31219} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1360.541615] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1360.542240] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.542418] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.542807] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1360.543070] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89ddd257-4e4c-4ec4-ad75-b84d8e32ed77 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1360.547309] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1360.547309] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bc908f-e7b6-078d-6e59-1159629735dc" [ 1360.547309] env[61907]: _type = "Task" [ 1360.547309] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1360.554334] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]52bc908f-e7b6-078d-6e59-1159629735dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1360.690733] env[61907]: DEBUG nova.compute.manager [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1360.690799] env[61907]: DEBUG nova.compute.manager [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing instance network info cache due to event network-changed-4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1360.691032] env[61907]: DEBUG oslo_concurrency.lockutils [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] Acquiring lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.691193] env[61907]: DEBUG oslo_concurrency.lockutils [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] Acquired lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.691364] env[61907]: DEBUG nova.network.neutron [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Refreshing network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1361.057281] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.057666] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Processing image aefba674-8a3f-4112-8a55-f47a5043daee {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1361.057776] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.057934] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquired lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.058137] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1361.058398] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-851a00fd-5efc-44e8-95f8-a4d6b5f9ae1d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.075033] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1361.075233] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1361.075918] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-420cc8ae-fbd4-4996-82f6-c929dbdcf9c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.080622] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1361.080622] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5260f700-b6d0-bfbc-f627-d00709020407" [ 1361.080622] env[61907]: _type = "Task" [ 1361.080622] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.087692] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5260f700-b6d0-bfbc-f627-d00709020407, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.382429] env[61907]: DEBUG nova.network.neutron [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updated VIF entry in instance network info cache for port 4ff33968-4c0e-40ce-97bf-32051f27be74. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1361.382830] env[61907]: DEBUG nova.network.neutron [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [{"id": "4ff33968-4c0e-40ce-97bf-32051f27be74", "address": "fa:16:3e:5c:61:6e", "network": {"id": "943e0833-d8f7-44c4-8668-7a63f67402f1", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1462305301-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.239", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f851fef407334ba2919877cfd4865435", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ff33968-4c", "ovs_interfaceid": "4ff33968-4c0e-40ce-97bf-32051f27be74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.590280] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Preparing fetch location {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1361.590494] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Fetch image to [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946/OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946.vmdk {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1361.590685] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Downloading stream optimized image aefba674-8a3f-4112-8a55-f47a5043daee to [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946/OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946.vmdk on the data store datastore2 as vApp {{(pid=61907) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1361.590862] env[61907]: DEBUG nova.virt.vmwareapi.images [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Downloading image file data aefba674-8a3f-4112-8a55-f47a5043daee to the ESX as VM named 'OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946' {{(pid=61907) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1361.656228] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1361.656228] env[61907]: value = "resgroup-9" [ 1361.656228] env[61907]: _type = "ResourcePool" [ 1361.656228] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1361.656512] env[61907]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-ccfc76bc-fc6c-4d13-96d3-cae71daea99f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.676738] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease: (returnval){ [ 1361.676738] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525b3ba8-46f4-c2a9-e926-3494f585dd0b" [ 1361.676738] env[61907]: _type = "HttpNfcLease" [ 1361.676738] env[61907]: } obtained for vApp import into resource pool (val){ [ 1361.676738] env[61907]: value = "resgroup-9" [ 1361.676738] env[61907]: _type = "ResourcePool" [ 1361.676738] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1361.677026] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the lease: (returnval){ [ 1361.677026] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525b3ba8-46f4-c2a9-e926-3494f585dd0b" [ 1361.677026] env[61907]: _type = "HttpNfcLease" [ 1361.677026] env[61907]: } to be ready. {{(pid=61907) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1361.683625] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1361.683625] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525b3ba8-46f4-c2a9-e926-3494f585dd0b" [ 1361.683625] env[61907]: _type = "HttpNfcLease" [ 1361.683625] env[61907]: } is initializing. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1361.886340] env[61907]: DEBUG oslo_concurrency.lockutils [req-8c0c985f-5eb0-497b-9f55-722fb54a9a4d req-87b2aaa7-7524-4e67-84e4-452ce839be0a service nova] Releasing lock "refresh_cache-26db17bf-ac9c-403f-b4fd-040db331706b" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.184586] env[61907]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1362.184586] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525b3ba8-46f4-c2a9-e926-3494f585dd0b" [ 1362.184586] env[61907]: _type = "HttpNfcLease" [ 1362.184586] env[61907]: } is ready. {{(pid=61907) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1362.185038] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1362.185038] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]525b3ba8-46f4-c2a9-e926-3494f585dd0b" [ 1362.185038] env[61907]: _type = "HttpNfcLease" [ 1362.185038] env[61907]: }. {{(pid=61907) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1362.185623] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbdb321a-5672-4cee-ad70-47e4c729f5cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.193726] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk from lease info. {{(pid=61907) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1362.193908] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk. {{(pid=61907) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1362.257167] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b3460539-afd8-4e80-af2d-8cbee5a20492 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.393215] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Completed reading data from the image iterator. {{(pid=61907) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1363.393584] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1363.394474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11dda0b0-d8e9-44c3-b523-3316b162e7e8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.401250] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk is in state: ready. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1363.401464] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk. {{(pid=61907) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1363.401708] env[61907]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-3fad841b-7339-45e5-9994-4d0cf74693dc {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.570897] env[61907]: DEBUG oslo_vmware.rw_handles [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/525999c2-aa0c-cdde-c812-40ee7d216177/disk-0.vmdk. {{(pid=61907) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1363.571124] env[61907]: INFO nova.virt.vmwareapi.images [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Downloaded image file data aefba674-8a3f-4112-8a55-f47a5043daee [ 1363.571962] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f68cab-7e29-482e-b2d2-4cfc27ebe1cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.586998] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d6053aa-c4eb-4691-b7be-9cdf891ad887 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.607982] env[61907]: INFO nova.virt.vmwareapi.images [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] The imported VM was unregistered [ 1363.610428] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Caching image {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1363.610669] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Creating directory with path [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1363.610986] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c3b727c-2673-448d-89b2-52bf6fd49bf0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.628575] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Created directory with path [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1363.628802] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946/OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946.vmdk to [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk. {{(pid=61907) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1363.629066] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-aed485e2-6b32-443c-825b-8ff38eedf2a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.634993] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1363.634993] env[61907]: value = "task-1244565" [ 1363.634993] env[61907]: _type = "Task" [ 1363.634993] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.642389] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.147131] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.647407] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.148335] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.647446] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.149220] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244565, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.181855} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.149526] env[61907]: INFO nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946/OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946.vmdk to [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk. [ 1366.149682] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Cleaning up location [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1366.149847] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_7845e899-7277-42ab-bc09-4a1ad6536946 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1366.150117] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-678f4601-2771-4cb2-b23f-d0f763a78631 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.156242] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1366.156242] env[61907]: value = "task-1244566" [ 1366.156242] env[61907]: _type = "Task" [ 1366.156242] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.163055] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244566, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.666472] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034528} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.666777] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1366.666945] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Releasing lock "[datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1366.667214] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk to [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1366.667468] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c6de2c9-3e46-42fd-abfd-b3c9706a0220 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.674602] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1366.674602] env[61907]: value = "task-1244567" [ 1366.674602] env[61907]: _type = "Task" [ 1366.674602] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1366.681865] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.185660] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task} progress is 26%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1367.687377] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task} progress is 49%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.189256] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1368.689519] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.189852] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244567, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.045975} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1369.190163] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/aefba674-8a3f-4112-8a55-f47a5043daee/aefba674-8a3f-4112-8a55-f47a5043daee.vmdk to [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1369.190910] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-182a11be-b64d-4e91-81ad-901c146eba18 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.211570] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Reconfiguring VM instance instance-00000070 to attach disk [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1369.211795] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584af29b-f5f8-4c85-ab24-d9cb504d4df0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1369.229584] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1369.229584] env[61907]: value = "task-1244568" [ 1369.229584] env[61907]: _type = "Task" [ 1369.229584] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1369.236597] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244568, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1369.739538] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244568, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.241260] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244568, 'name': ReconfigVM_Task, 'duration_secs': 0.538077} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.241498] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Reconfigured VM instance instance-00000070 to attach disk [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b/26db17bf-ac9c-403f-b4fd-040db331706b.vmdk or device None with type streamOptimized {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1370.242123] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-918b26a5-e1b2-4927-8242-0d4d2d86f448 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.247614] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1370.247614] env[61907]: value = "task-1244569" [ 1370.247614] env[61907]: _type = "Task" [ 1370.247614] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.254557] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244569, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1370.757467] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244569, 'name': Rename_Task, 'duration_secs': 0.162854} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1370.757817] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1370.757970] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fec0e11-2bdd-4a60-a1a3-b446fc3fc55d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1370.763886] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1370.763886] env[61907]: value = "task-1244570" [ 1370.763886] env[61907]: _type = "Task" [ 1370.763886] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1370.770792] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.274641] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1371.775914] env[61907]: DEBUG oslo_vmware.api [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244570, 'name': PowerOnVM_Task, 'duration_secs': 0.994317} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1371.776306] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1371.872507] env[61907]: DEBUG nova.compute.manager [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1371.873460] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86fc9a6e-c03f-4b40-b844-4c6c0a61714a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1372.390607] env[61907]: DEBUG oslo_concurrency.lockutils [None req-a5ec6ab1-542a-4476-894a-5bccf69930db tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 19.482s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.952139] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.952549] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.952659] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1372.952888] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1372.953119] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1372.955368] env[61907]: INFO nova.compute.manager [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Terminating instance [ 1373.460058] env[61907]: DEBUG nova.compute.manager [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1373.460058] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1373.460693] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-096188df-6dc4-4974-bae8-8ee515bc43a3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.468521] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1373.468755] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4d1cbc3d-7664-4892-901c-4920285fbb3f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1373.475052] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1373.475052] env[61907]: value = "task-1244571" [ 1373.475052] env[61907]: _type = "Task" [ 1373.475052] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1373.482174] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1373.984045] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244571, 'name': PowerOffVM_Task, 'duration_secs': 0.185931} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1373.984427] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1373.984512] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1373.984756] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1821d35f-7030-4c88-95df-bae6ed2193c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.042137] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1374.042335] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1374.042522] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleting the datastore file [datastore2] 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1374.042785] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-78377cea-3bae-4c7f-af4e-33cff1416756 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.048766] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for the task: (returnval){ [ 1374.048766] env[61907]: value = "task-1244573" [ 1374.048766] env[61907]: _type = "Task" [ 1374.048766] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1374.055957] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1374.557892] env[61907]: DEBUG oslo_vmware.api [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Task: {'id': task-1244573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144196} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1374.558169] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1374.558364] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1374.558556] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1374.558725] env[61907]: INFO nova.compute.manager [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1374.559017] env[61907]: DEBUG oslo.service.loopingcall [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1374.559249] env[61907]: DEBUG nova.compute.manager [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1374.559346] env[61907]: DEBUG nova.network.neutron [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1374.990798] env[61907]: DEBUG nova.compute.manager [req-8a0ae7f1-170b-481c-8fcd-e84aaf966271 req-d268e05b-6b46-492c-813a-9ae7304c7674 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Received event network-vif-deleted-4ff33968-4c0e-40ce-97bf-32051f27be74 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1374.991167] env[61907]: INFO nova.compute.manager [req-8a0ae7f1-170b-481c-8fcd-e84aaf966271 req-d268e05b-6b46-492c-813a-9ae7304c7674 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Neutron deleted interface 4ff33968-4c0e-40ce-97bf-32051f27be74; detaching it from the instance and deleting it from the info cache [ 1374.991167] env[61907]: DEBUG nova.network.neutron [req-8a0ae7f1-170b-481c-8fcd-e84aaf966271 req-d268e05b-6b46-492c-813a-9ae7304c7674 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1375.469931] env[61907]: DEBUG nova.network.neutron [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1375.494252] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f424fc0-5c69-485e-9c19-1281a50ecfeb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.504265] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac39e49-445f-44ec-aa4e-919a7e26e2cf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.527684] env[61907]: DEBUG nova.compute.manager [req-8a0ae7f1-170b-481c-8fcd-e84aaf966271 req-d268e05b-6b46-492c-813a-9ae7304c7674 service nova] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Detach interface failed, port_id=4ff33968-4c0e-40ce-97bf-32051f27be74, reason: Instance 26db17bf-ac9c-403f-b4fd-040db331706b could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1375.975731] env[61907]: INFO nova.compute.manager [-] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Took 1.42 seconds to deallocate network for instance. [ 1376.482255] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1376.482640] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1376.482855] env[61907]: DEBUG nova.objects.instance [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lazy-loading 'resources' on Instance uuid 26db17bf-ac9c-403f-b4fd-040db331706b {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1377.028408] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2334ac75-f22f-4018-b9d6-56ef48ea38f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.035977] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4347c0-75a7-4c58-9232-4499ccb89ffe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.064510] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37686eef-49e6-4fb1-b99a-44d1e20bacf7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.071464] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f0699ad-d696-4cd2-af3a-dea3b67828b2 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1377.084370] env[61907]: DEBUG nova.compute.provider_tree [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1377.587196] env[61907]: DEBUG nova.scheduler.client.report [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1378.092725] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.610s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1378.111703] env[61907]: INFO nova.scheduler.client.report [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Deleted allocations for instance 26db17bf-ac9c-403f-b4fd-040db331706b [ 1378.619189] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f2aa19fe-e102-48ee-8c9a-6536605df34c tempest-ServerActionsTestOtherB-1634310954 tempest-ServerActionsTestOtherB-1634310954-project-member] Lock "26db17bf-ac9c-403f-b4fd-040db331706b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.667s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1391.514701] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1391.515126] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances with incomplete migration {{(pid=61907) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 1399.011168] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1401.513894] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1402.513889] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1403.514341] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1403.514645] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1404.017506] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Didn't find any instances for network info cache update. {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10357}} [ 1404.017720] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.017928] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.018078] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 1404.524741] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] There are 12 instances to clean {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 1404.525168] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 26db17bf-ac9c-403f-b4fd-040db331706b] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1405.028671] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 97e5b53f-557e-441a-9097-ffdb5a86ad38] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1405.531791] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: ad8a7d32-6ef3-44d2-9617-49335af29ae2] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1406.035035] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f1dc869a-473d-4d7e-bf4b-c46498c55c56] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1406.538740] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 868a8e82-db37-4a6d-86e7-64811c3c4840] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1407.041958] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: bc6d118d-41c7-4a7a-9b75-14919844899a] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1407.546026] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 0e9c2580-cf31-45fd-b5a7-ee13291056aa] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1408.049683] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 32fc7dba-6a76-4f0b-bca2-204e3bc228de] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1408.553159] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 378b5c99-5dbc-4657-8504-58d20dc6c9ea] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1409.056462] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: b2c5689e-f930-4548-b061-fd2974d44575] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1409.560058] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f7959e23-8879-4285-93da-804dd3115f65] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1410.063921] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: f72373bf-80c3-49b9-af8d-bc6b1ec2d6fa] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1410.567468] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.565688] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.565998] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.566218] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1411.566366] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1411.566516] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1412.069462] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1412.069843] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1412.069946] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1412.070102] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1412.071163] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fffa7bd-6efd-40f6-884d-4e71c11ba912 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.079371] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5483c1ef-3e79-4177-910a-486d6581f56b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.092447] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5065a595-75b3-40fc-ac7b-ef9e34144866 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.098671] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcd2e93-18a9-4908-8da8-95fb5fdfa02d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1412.127301] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181303MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1412.127485] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1412.127678] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1413.274377] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c310aa25-4daa-474b-a801-295269d1b5f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1413.274680] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1413.274811] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1413.301331] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132b8f0a-cb60-4140-aeda-2edf7fa1d452 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.308685] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188b7d91-f120-4c10-941e-40a8b4b830d5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.337450] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729cfe8e-1bd7-4afc-9f15-007be32d3c4d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.343956] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22aa68ca-8979-499a-a875-21d4ebb5e233 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.356255] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1413.859552] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1414.364711] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1414.365105] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.237s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1461.308211] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1461.514019] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.515608] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1463.515940] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1463.515940] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1464.131071] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1464.131198] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1464.131371] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1464.131543] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid c310aa25-4daa-474b-a801-295269d1b5f0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1465.870658] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [{"id": "79dafada-2206-4800-8dd1-07aca1f973c4", "address": "fa:16:3e:6a:5d:60", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.234", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79dafada-22", "ovs_interfaceid": "79dafada-2206-4800-8dd1-07aca1f973c4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1466.373792] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-c310aa25-4daa-474b-a801-295269d1b5f0" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1466.374036] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1466.374286] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1466.374446] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1466.374604] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1466.374736] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1466.513968] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1467.019410] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1467.019771] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1467.019771] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1467.522676] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1467.522931] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1467.523112] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1467.523269] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1467.524191] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9416726-3aea-4d9d-ae7d-894a7778d124 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.532274] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9100f10-3704-4f5b-9bd1-62e122ae7226 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.545586] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd289089-e4e2-483a-b6c3-b6037508c4b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.551453] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60adb2d8-1a42-4248-b6d2-fe3acf8f6d39 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1467.579500] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181348MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1467.579648] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1467.579840] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1468.604264] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance c310aa25-4daa-474b-a801-295269d1b5f0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1468.604264] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1468.604264] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1468.627087] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cea1c57-c486-4189-8cd6-b359c0b5ed49 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.634361] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94275c0d-e27b-464e-9a61-6433b6d8e98d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.662827] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7be414f1-7506-461e-8af9-5d160a428e23 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.669176] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0016c8-bc19-4774-9ed3-158555df4a68 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1468.682719] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1469.185803] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1469.187097] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1469.187295] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.607s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1503.409064] env[61907]: INFO nova.compute.manager [None req-dc78351d-bd49-4bf3-bb39-ecdc3afadfa1 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Get console output [ 1503.409381] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-dc78351d-bd49-4bf3-bb39-ecdc3afadfa1 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] The console log is missing. Check your VSPC configuration [ 1504.490459] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "c310aa25-4daa-474b-a801-295269d1b5f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1504.490830] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1504.490951] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1504.491205] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1504.491389] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1504.494859] env[61907]: INFO nova.compute.manager [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Terminating instance [ 1504.998703] env[61907]: DEBUG nova.compute.manager [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1504.998964] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1504.999897] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89967906-3aa2-4f9b-8214-57462550f60d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.007560] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1505.007798] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5e17ef9-081b-48f5-a959-c148e58ef42d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.014170] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1505.014170] env[61907]: value = "task-1244575" [ 1505.014170] env[61907]: _type = "Task" [ 1505.014170] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1505.021661] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244575, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1505.524383] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244575, 'name': PowerOffVM_Task, 'duration_secs': 0.200363} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1505.524684] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1505.524818] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1505.525076] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9d8ffb0-8437-4deb-a4ce-5c1ea79bdabd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.585107] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1505.585385] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Deleting contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1505.585659] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleting the datastore file [datastore1] c310aa25-4daa-474b-a801-295269d1b5f0 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1505.585926] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6343c4b-782a-4e52-ae67-bb9f1067468b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1505.593618] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1505.593618] env[61907]: value = "task-1244577" [ 1505.593618] env[61907]: _type = "Task" [ 1505.593618] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1505.601497] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244577, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1506.103946] env[61907]: DEBUG oslo_vmware.api [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244577, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128293} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1506.104189] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1506.104382] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Deleted contents of the VM from datastore datastore1 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1506.104565] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1506.104742] env[61907]: INFO nova.compute.manager [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1506.104998] env[61907]: DEBUG oslo.service.loopingcall [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1506.105214] env[61907]: DEBUG nova.compute.manager [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1506.105310] env[61907]: DEBUG nova.network.neutron [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1506.556653] env[61907]: DEBUG nova.compute.manager [req-90d69868-2a8a-43eb-847e-594e8e55f843 req-234dcefe-930e-4134-b94e-6f322c924f45 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Received event network-vif-deleted-79dafada-2206-4800-8dd1-07aca1f973c4 {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1506.556883] env[61907]: INFO nova.compute.manager [req-90d69868-2a8a-43eb-847e-594e8e55f843 req-234dcefe-930e-4134-b94e-6f322c924f45 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Neutron deleted interface 79dafada-2206-4800-8dd1-07aca1f973c4; detaching it from the instance and deleting it from the info cache [ 1506.557045] env[61907]: DEBUG nova.network.neutron [req-90d69868-2a8a-43eb-847e-594e8e55f843 req-234dcefe-930e-4134-b94e-6f322c924f45 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1507.004601] env[61907]: DEBUG nova.network.neutron [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1507.059169] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3809344b-38f1-4832-82e7-b7c03725f15d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1507.070995] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-863038fb-ca64-4fe2-b5e8-b7f596852b2a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1507.092343] env[61907]: DEBUG nova.compute.manager [req-90d69868-2a8a-43eb-847e-594e8e55f843 req-234dcefe-930e-4134-b94e-6f322c924f45 service nova] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Detach interface failed, port_id=79dafada-2206-4800-8dd1-07aca1f973c4, reason: Instance c310aa25-4daa-474b-a801-295269d1b5f0 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1507.507298] env[61907]: INFO nova.compute.manager [-] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Took 1.40 seconds to deallocate network for instance. [ 1508.014384] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1508.014747] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1508.014885] env[61907]: DEBUG nova.objects.instance [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lazy-loading 'resources' on Instance uuid c310aa25-4daa-474b-a801-295269d1b5f0 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1508.550039] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb73117-5621-4ffe-bd1f-909d61fb9ab6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.557817] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a449c3-38d7-49bb-907a-931277a0d4ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.588829] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-938d6da2-88df-413d-a95c-4f4911f489eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.596584] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65fba08-0220-4c1b-b2a7-a4bafe3fcd42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1508.609789] env[61907]: DEBUG nova.compute.provider_tree [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1509.113414] env[61907]: DEBUG nova.scheduler.client.report [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1509.619016] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.604s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1509.642819] env[61907]: INFO nova.scheduler.client.report [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted allocations for instance c310aa25-4daa-474b-a801-295269d1b5f0 [ 1510.151824] env[61907]: DEBUG oslo_concurrency.lockutils [None req-5b29a634-03dd-40ff-a8a8-43862b0ed892 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "c310aa25-4daa-474b-a801-295269d1b5f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.661s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1512.560410] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "5261031b-de16-432b-87f8-68c735995ce4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1512.560741] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1513.063641] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Starting instance... {{(pid=61907) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2438}} [ 1513.588386] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1513.588648] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1513.590253] env[61907]: INFO nova.compute.claims [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1514.627673] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a87def-be97-4c12-aa99-963ce1c50708 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.635115] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60230e2-a6e0-4bda-8132-a0e9afc00863 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.664325] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad5272e-f5bd-4046-9c43-d9fe59f279b1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.671510] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abdd46a-7809-480c-88ac-c49c2a4967f0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1514.684195] env[61907]: DEBUG nova.compute.provider_tree [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1515.187104] env[61907]: DEBUG nova.scheduler.client.report [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1515.692662] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.104s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1515.693298] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Start building networks asynchronously for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2835}} [ 1516.199425] env[61907]: DEBUG nova.compute.utils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Using /dev/sd instead of None {{(pid=61907) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1516.201525] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Allocating IP information in the background. {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1987}} [ 1516.201525] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] allocate_for_instance() {{(pid=61907) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1516.250571] env[61907]: DEBUG nova.policy [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b4762c5f79e48fb8a4f9701e5c59ff3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8c30836c7053443aaab6e1dbf7783259', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61907) authorize /opt/stack/nova/nova/policy.py:201}} [ 1516.524185] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Successfully created port: 7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1516.705210] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Start building block device mappings for instance. {{(pid=61907) _build_resources /opt/stack/nova/nova/compute/manager.py:2870}} [ 1517.716685] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Start spawning the instance on the hypervisor. {{(pid=61907) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2644}} [ 1517.745839] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-10T13:05:31Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-10T13:05:13Z,direct_url=,disk_format='vmdk',id=fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='bd52a0efc1a64b4db183360d593bda95',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-10T13:05:14Z,virtual_size=,visibility=), allow threads: False {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1517.746116] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1517.746289] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image limits 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1517.746483] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Flavor pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1517.746635] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Image pref 0:0:0 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1517.746790] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61907) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1517.747116] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1517.747305] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1517.747482] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Got 1 possible topologies {{(pid=61907) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1517.747688] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1517.747887] env[61907]: DEBUG nova.virt.hardware [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61907) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1517.748748] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b440043-717e-4384-8372-e32bab7aa93f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.756615] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-006583bb-3c88-402b-a000-2839b94e85c7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1517.884438] env[61907]: DEBUG nova.compute.manager [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Received event network-vif-plugged-7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1517.884836] env[61907]: DEBUG oslo_concurrency.lockutils [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] Acquiring lock "5261031b-de16-432b-87f8-68c735995ce4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1517.884969] env[61907]: DEBUG oslo_concurrency.lockutils [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] Lock "5261031b-de16-432b-87f8-68c735995ce4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1517.885058] env[61907]: DEBUG oslo_concurrency.lockutils [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] Lock "5261031b-de16-432b-87f8-68c735995ce4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1517.885239] env[61907]: DEBUG nova.compute.manager [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] No waiting events found dispatching network-vif-plugged-7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) pop_instance_event /opt/stack/nova/nova/compute/manager.py:322}} [ 1517.885409] env[61907]: WARNING nova.compute.manager [req-78e1326a-b596-412d-b4c4-4e767768e1d0 req-a6c587e5-b358-4841-bf44-be886d90f08b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Received unexpected event network-vif-plugged-7528f86d-8b84-4a22-970f-f3db5232a72b for instance with vm_state building and task_state spawning. [ 1517.971336] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Successfully updated port: 7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1518.474158] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1518.474320] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1518.474474] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Building network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1519.005669] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Instance cache missing network info. {{(pid=61907) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1519.125678] env[61907]: DEBUG nova.network.neutron [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1519.628461] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1519.628798] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Instance network_info: |[{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61907) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2002}} [ 1519.629307] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bc:81:70', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7528f86d-8b84-4a22-970f-f3db5232a72b', 'vif_model': 'vmxnet3'}] {{(pid=61907) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1519.636784] env[61907]: DEBUG oslo.service.loopingcall [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1519.636995] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Creating VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1519.637618] env[61907]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49ced668-eb1f-44c6-a688-0473f69384b5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1519.657858] env[61907]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1519.657858] env[61907]: value = "task-1244578" [ 1519.657858] env[61907]: _type = "Task" [ 1519.657858] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1519.665377] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244578, 'name': CreateVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1519.910542] env[61907]: DEBUG nova.compute.manager [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Received event network-changed-7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1519.910715] env[61907]: DEBUG nova.compute.manager [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Refreshing instance network info cache due to event network-changed-7528f86d-8b84-4a22-970f-f3db5232a72b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1519.910945] env[61907]: DEBUG oslo_concurrency.lockutils [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1519.911110] env[61907]: DEBUG oslo_concurrency.lockutils [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1519.911329] env[61907]: DEBUG nova.network.neutron [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Refreshing network info cache for port 7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1520.168353] env[61907]: DEBUG oslo_vmware.api [-] Task: {'id': task-1244578, 'name': CreateVM_Task, 'duration_secs': 0.293961} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1520.168712] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Created VM on the ESX host {{(pid=61907) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1520.175786] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1520.175963] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1520.176319] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1520.176574] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a46df5b7-b2e7-46c9-aaca-6204dbe197e1 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.181826] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1520.181826] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225ba53-78a5-70b9-f4b8-036edf8c575f" [ 1520.181826] env[61907]: _type = "Task" [ 1520.181826] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1520.189652] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225ba53-78a5-70b9-f4b8-036edf8c575f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1520.593957] env[61907]: DEBUG nova.network.neutron [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated VIF entry in instance network info cache for port 7528f86d-8b84-4a22-970f-f3db5232a72b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1520.594351] env[61907]: DEBUG nova.network.neutron [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1520.691708] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]5225ba53-78a5-70b9-f4b8-036edf8c575f, 'name': SearchDatastore_Task, 'duration_secs': 0.010854} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1520.692037] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1520.692315] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Processing image fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57 {{(pid=61907) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1520.692576] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1520.692732] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1520.692917] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1520.693203] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c2502a9f-252d-4aa2-b506-7dccc680d8eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.701470] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61907) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1520.701692] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61907) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1520.702479] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95fb866f-3b1a-464d-ac82-5a75c3c501bf {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1520.707560] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1520.707560] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d0874-bebd-0e18-33fe-d82796cce28d" [ 1520.707560] env[61907]: _type = "Task" [ 1520.707560] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1520.715326] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d0874-bebd-0e18-33fe-d82796cce28d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1521.097207] env[61907]: DEBUG oslo_concurrency.lockutils [req-15e01467-c865-4917-8ae5-6f7e93aff79c req-71f81114-83d1-4229-9561-99f25db41571 service nova] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1521.217674] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]524d0874-bebd-0e18-33fe-d82796cce28d, 'name': SearchDatastore_Task, 'duration_secs': 0.009607} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1521.218473] env[61907]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-806adc1d-0f8a-4feb-b0a5-76fd58500e4a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.223579] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1521.223579] env[61907]: value = "session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529d212b-656a-5cda-f949-6de2b294d027" [ 1521.223579] env[61907]: _type = "Task" [ 1521.223579] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1521.230484] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529d212b-656a-5cda-f949-6de2b294d027, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1521.734475] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': session[52c13d70-3256-a0f8-345f-64ed8bab5b88]529d212b-656a-5cda-f949-6de2b294d027, 'name': SearchDatastore_Task, 'duration_secs': 0.011484} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1521.734750] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1521.735014] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5261031b-de16-432b-87f8-68c735995ce4/5261031b-de16-432b-87f8-68c735995ce4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1521.735285] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-73a4b622-27c8-465a-868b-5062f72e2aa5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1521.741906] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1521.741906] env[61907]: value = "task-1244579" [ 1521.741906] env[61907]: _type = "Task" [ 1521.741906] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1521.749875] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.252021] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.420138} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1522.252387] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57/fe0a0c14-6a2f-465c-80aa-aecd8c9e4d57.vmdk to [datastore2] 5261031b-de16-432b-87f8-68c735995ce4/5261031b-de16-432b-87f8-68c735995ce4.vmdk {{(pid=61907) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1522.252479] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Extending root virtual disk to 1048576 {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1522.252686] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-804c9fc1-7489-40da-a046-56bace2d4b97 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.259265] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1522.259265] env[61907]: value = "task-1244580" [ 1522.259265] env[61907]: _type = "Task" [ 1522.259265] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1522.266130] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1522.769289] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061131} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1522.769562] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Extended root virtual disk {{(pid=61907) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1522.770332] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7ab4300-e5ac-4c9c-9352-591c1b563dfe {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.791058] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Reconfiguring VM instance instance-00000071 to attach disk [datastore2] 5261031b-de16-432b-87f8-68c735995ce4/5261031b-de16-432b-87f8-68c735995ce4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1522.791305] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45bc336d-a1f4-416e-8903-188a3ee09518 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1522.809916] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1522.809916] env[61907]: value = "task-1244581" [ 1522.809916] env[61907]: _type = "Task" [ 1522.809916] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1522.817687] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244581, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1523.181729] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.181973] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1523.319293] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244581, 'name': ReconfigVM_Task, 'duration_secs': 0.263491} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1523.319672] env[61907]: DEBUG nova.virt.vmwareapi.volumeops [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Reconfigured VM instance instance-00000071 to attach disk [datastore2] 5261031b-de16-432b-87f8-68c735995ce4/5261031b-de16-432b-87f8-68c735995ce4.vmdk or device None with type sparse {{(pid=61907) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1523.320212] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e2a8c53-9d93-4390-a949-ffd04b99c1c8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1523.326215] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1523.326215] env[61907]: value = "task-1244582" [ 1523.326215] env[61907]: _type = "Task" [ 1523.326215] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1523.334286] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244582, 'name': Rename_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1523.835592] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244582, 'name': Rename_Task, 'duration_secs': 0.137158} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1523.835868] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Powering on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1523.836127] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c8c2888-e798-456f-9785-daf4a199f993 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1523.841586] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1523.841586] env[61907]: value = "task-1244583" [ 1523.841586] env[61907]: _type = "Task" [ 1523.841586] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1523.848396] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1524.351671] env[61907]: DEBUG oslo_vmware.api [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244583, 'name': PowerOnVM_Task, 'duration_secs': 0.426112} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1524.351945] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Powered on the VM {{(pid=61907) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1524.352172] env[61907]: INFO nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Took 6.64 seconds to spawn the instance on the hypervisor. [ 1524.352356] env[61907]: DEBUG nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Checking state {{(pid=61907) _get_power_state /opt/stack/nova/nova/compute/manager.py:1797}} [ 1524.353139] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a87b732-e3a6-41d2-ad4d-51fc12b1f67b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1524.513759] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1524.870444] env[61907]: INFO nova.compute.manager [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Took 11.30 seconds to build instance. [ 1525.372995] env[61907]: DEBUG oslo_concurrency.lockutils [None req-290d3100-169b-4c3b-bbbc-1f63b9d73fef tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.812s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1525.513243] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1525.513421] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1525.513531] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1525.797721] env[61907]: DEBUG nova.compute.manager [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Received event network-changed-7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1525.797940] env[61907]: DEBUG nova.compute.manager [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Refreshing instance network info cache due to event network-changed-7528f86d-8b84-4a22-970f-f3db5232a72b. {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11479}} [ 1525.798182] env[61907]: DEBUG oslo_concurrency.lockutils [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1525.798330] env[61907]: DEBUG oslo_concurrency.lockutils [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1525.798496] env[61907]: DEBUG nova.network.neutron [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Refreshing network info cache for port 7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1526.044948] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1526.507187] env[61907]: DEBUG nova.network.neutron [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated VIF entry in instance network info cache for port 7528f86d-8b84-4a22-970f-f3db5232a72b. {{(pid=61907) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1526.507569] env[61907]: DEBUG nova.network.neutron [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1527.010875] env[61907]: DEBUG oslo_concurrency.lockutils [req-d3363a88-ded6-4b6f-8a63-41d9563485b3 req-14dc9980-028c-47fd-9028-9f6f8cac517b service nova] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1527.011360] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1527.011572] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1527.011736] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1528.714672] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1529.217637] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1529.217914] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1529.218177] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.218380] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.218581] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.218779] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.218931] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1529.219129] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1529.722032] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1529.722432] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1529.722488] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1529.722648] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1529.723595] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44105b6a-5911-492f-a435-371cdf3cdbc7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.731655] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605277c3-d625-44af-9d48-c2cc826719eb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.744839] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9568de-b02f-4607-9a6e-40c77c72e416 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.750497] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ce82b9-ec13-4b76-933f-e923f33c5e78 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1529.778171] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181506MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1529.778313] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1529.778499] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1530.802362] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5261031b-de16-432b-87f8-68c735995ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1530.802629] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1530.802780] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1530.827612] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b204180-4735-4056-831e-38826f082036 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.835211] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d276ee4-eae7-4dd7-910f-8ce9117d55a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.865468] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3721c0-f626-4a6c-b5b5-5fde3c2c779f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.872567] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e39e29d-018a-43b8-9eba-3412d772a680 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1530.885579] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1531.388483] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1531.893765] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1531.894149] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1586.190323] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1586.190730] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1586.513312] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1586.513528] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1586.513599] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1587.051297] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1587.051523] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1587.051755] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1587.051982] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1588.756211] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1589.258607] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1589.258849] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1589.259075] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1589.259242] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1589.259395] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1589.259524] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1589.514080] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1590.019258] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1590.019612] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1590.513935] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1591.016940] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1591.017203] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1591.017347] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1591.017507] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1591.018450] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-785b07d6-e8f8-4c25-bc58-fe9d28006e35 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.027141] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ecd616-6309-4ef4-acbc-05337e37313b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.040638] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c709d61d-dff6-45c2-9b5c-682526c9f772 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.046771] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3079a00-4cd8-473d-8d56-a528393141be {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1591.074660] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181363MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1591.074790] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1591.074976] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1592.095649] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5261031b-de16-432b-87f8-68c735995ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1592.095928] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1592.096102] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1592.120609] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7ce76c-4179-4081-9cce-52d14f6232c0 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.128045] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1282e19-939b-4844-9d13-44243d3231c9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.156391] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fa524fe-b9b9-4872-b369-6062539a9bd5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.162972] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe4b6b3-85d9-4dcf-89c6-df606d5e3240 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1592.175331] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1592.678123] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1592.679416] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1592.679616] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.605s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1644.674356] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1644.674744] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1646.514835] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1647.513773] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1647.513980] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1647.514129] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1648.045145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1648.045356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1648.045440] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1648.045588] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1649.751875] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1650.254126] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1650.254351] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1650.254547] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1650.255892] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1650.255892] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1650.514390] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1651.513944] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1651.514318] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1652.017578] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1652.017817] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1652.018025] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1652.018180] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1652.019107] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47ca3ed6-fe5b-429b-ac38-713a45c662f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1652.027525] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b6f424-18d0-4831-84e4-9a410675e07f {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1652.040948] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4a07f6-1a7b-4433-bbe0-a7549d015fa5 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1652.046980] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edf9137-939b-4020-990d-b36d276d59a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1652.075669] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181366MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1652.075772] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1652.075946] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1653.096173] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5261031b-de16-432b-87f8-68c735995ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1653.096438] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1653.096528] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1653.113317] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing inventories for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:819}} [ 1653.125335] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating ProviderTree inventory for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:783}} [ 1653.125517] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Updating inventory in ProviderTree for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1653.135997] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing aggregate associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, aggregates: None {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:828}} [ 1653.153804] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Refreshing trait associations for resource provider 91bca385-a423-4ca4-9da0-aeb4615e22d3, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=61907) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:840}} [ 1653.175746] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14a72056-a1df-495c-b5f6-692493cc85d7 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.183044] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc73bbd-9a8b-4595-a8ca-f6c67bda9864 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.211462] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4342ab83-6424-409d-9a3f-ea83c9d8a0b3 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.218161] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c6af83-cd36-4ea9-93f2-a53183bf07d9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1653.231929] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1653.734851] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1653.736156] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1653.736356] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.660s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1704.514574] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1704.515040] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1704.515040] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11571}} [ 1705.018736] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] There are 1 instances to clean {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11580}} [ 1705.018959] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: c310aa25-4daa-474b-a801-295269d1b5f0] Instance has had 0 of 5 cleanup attempts {{(pid=61907) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11584}} [ 1705.522426] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1705.522761] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Cleaning up deleted instances with incomplete migration {{(pid=61907) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11609}} [ 1708.024397] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1708.513756] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1708.514043] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1708.514043] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1709.042894] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1709.043190] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1709.043230] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1709.043369] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1710.759996] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1711.262684] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1711.262989] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1711.263229] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1711.263397] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1711.263556] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1711.513733] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1712.019124] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1712.019514] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1712.019514] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1712.019644] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1712.522936] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1712.523215] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1712.523388] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1712.523546] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1712.524496] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-357ae1f5-b3fc-49e4-a2b9-93c6881e0f54 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.532735] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e944f70-fe06-46f7-bc5f-124500554506 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.546651] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249ee41e-523f-4b01-aa05-2d7784195df9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.552474] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea2f70b-a709-4bfe-8101-e495875c93f6 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1712.579820] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181371MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1712.579970] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1712.580145] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1713.647999] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5261031b-de16-432b-87f8-68c735995ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1713.648258] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1713.648376] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1713.673251] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d05a5c2-3ea6-4147-a93f-6d746b07520b {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1713.680430] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1bc51d5-d400-4271-89cf-bef73103340a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1713.710874] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec8331e1-4d88-4385-9d25-50498970fef4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1713.717573] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292926d4-bfa2-44b8-81fc-68aee5bde5a8 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1713.729817] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1714.232392] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1714.233726] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1714.233915] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.654s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1724.516740] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1765.010269] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1766.514392] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1768.514096] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1769.514229] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1769.514704] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Starting heal instance info cache {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10271}} [ 1769.514704] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Rebuilding the list of instances to heal {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10275}} [ 1770.083650] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1770.083816] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquired lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1770.083966] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Forcefully refreshing network info cache for instance {{(pid=61907) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1770.084146] env[61907]: DEBUG nova.objects.instance [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lazy-loading 'info_cache' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1771.789806] env[61907]: DEBUG nova.network.neutron [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [{"id": "7528f86d-8b84-4a22-970f-f3db5232a72b", "address": "fa:16:3e:bc:81:70", "network": {"id": "cabd88f6-2ff9-4267-9ce4-65910ad88356", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1476113417-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8c30836c7053443aaab6e1dbf7783259", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7528f86d-8b", "ovs_interfaceid": "7528f86d-8b84-4a22-970f-f3db5232a72b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1772.292339] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Releasing lock "refresh_cache-5261031b-de16-432b-87f8-68c735995ce4" {{(pid=61907) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1772.292594] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updated the network info_cache for instance {{(pid=61907) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10342}} [ 1772.292805] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1772.293131] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1772.293131] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61907) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10890}} [ 1772.513497] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1772.513746] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1772.513898] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager.update_available_resource {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1773.018136] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1773.018635] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1773.018635] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1773.018719] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61907) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1773.019637] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba063d45-7199-4748-9bf6-9067ef580517 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1773.027865] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f99e1f-b36a-423d-84cd-889cb80e6218 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1773.041312] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-656a4f43-e114-4160-a642-596e99356bf9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1773.047428] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b92a7a-9e94-4a2b-b796-2c2a3d509207 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1773.075650] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181370MB free_disk=148GB free_vcpus=48 pci_devices=None {{(pid=61907) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1773.075804] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1773.075976] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1774.097034] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Instance 5261031b-de16-432b-87f8-68c735995ce4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61907) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1774.097317] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1774.097432] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=61907) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1774.122564] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f81431-d4a8-4771-b181-fbb6591ec3ed {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1774.130240] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa286e4-f77c-40c4-862c-67dbe45e9a6a {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1774.159275] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38291051-867b-4362-a98a-6fea0c854f42 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1774.165802] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff816097-cf54-4512-a98c-ded748e2c1f9 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1774.178075] env[61907]: DEBUG nova.compute.provider_tree [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1774.681349] env[61907]: DEBUG nova.scheduler.client.report [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1774.682619] env[61907]: DEBUG nova.compute.resource_tracker [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61907) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1774.682803] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.607s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1774.864212] env[61907]: DEBUG oslo_service.periodic_task [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Running periodic task ComputeManager._sync_power_states {{(pid=61907) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1775.367547] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Getting list of instances from cluster (obj){ [ 1775.367547] env[61907]: value = "domain-c8" [ 1775.367547] env[61907]: _type = "ClusterComputeResource" [ 1775.367547] env[61907]: } {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1775.368985] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e3154d-a662-435e-96d9-b618a737c747 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1775.378025] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Got total of 1 instances {{(pid=61907) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1775.378198] env[61907]: DEBUG nova.compute.manager [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Triggering sync for uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10681}} [ 1775.378530] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Acquiring lock "5261031b-de16-432b-87f8-68c735995ce4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1775.378749] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "5261031b-de16-432b-87f8-68c735995ce4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1775.379538] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f639aae-04a8-43eb-837f-d4026b8949f4 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1775.888717] env[61907]: DEBUG oslo_concurrency.lockutils [None req-f7dff67b-7cda-4e2f-b0f4-3c739ad4365b None None] Lock "5261031b-de16-432b-87f8-68c735995ce4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1776.560017] env[61907]: INFO nova.compute.manager [None req-1749f5d4-29aa-4180-8915-b15f9e293f34 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Get console output [ 1776.560299] env[61907]: WARNING nova.virt.vmwareapi.driver [None req-1749f5d4-29aa-4180-8915-b15f9e293f34 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] The console log is missing. Check your VSPC configuration [ 1778.138323] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "5261031b-de16-432b-87f8-68c735995ce4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1778.138708] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1778.138853] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "5261031b-de16-432b-87f8-68c735995ce4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1778.139059] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1778.139241] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1778.142646] env[61907]: INFO nova.compute.manager [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Terminating instance [ 1778.646233] env[61907]: DEBUG nova.compute.manager [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Start destroying the instance on the hypervisor. {{(pid=61907) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3163}} [ 1778.646528] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Destroying instance {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1778.647457] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98ff54e-7dac-453c-849f-9944d4ea781d {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1778.655230] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Powering off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1778.655477] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e5f5b586-5fd0-4f0c-b07b-4e5490f066db {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1778.661152] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1778.661152] env[61907]: value = "task-1244584" [ 1778.661152] env[61907]: _type = "Task" [ 1778.661152] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1778.668820] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1779.171811] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244584, 'name': PowerOffVM_Task, 'duration_secs': 0.178775} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1779.171811] env[61907]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Powered off the VM {{(pid=61907) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1779.171811] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Unregistering the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1779.171811] env[61907]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-764a49bf-cbe5-4e34-a98c-10ce3057b18e {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1779.232899] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Unregistered the VM {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1779.233140] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Deleting contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1779.233330] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleting the datastore file [datastore2] 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1779.233593] env[61907]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4a7bd09f-7d3f-4f2b-8952-d958fa1906bb {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1779.240806] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for the task: (returnval){ [ 1779.240806] env[61907]: value = "task-1244586" [ 1779.240806] env[61907]: _type = "Task" [ 1779.240806] env[61907]: } to complete. {{(pid=61907) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1779.247961] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1779.751204] env[61907]: DEBUG oslo_vmware.api [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Task: {'id': task-1244586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146695} completed successfully. {{(pid=61907) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1779.751486] env[61907]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted the datastore file {{(pid=61907) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1779.751703] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Deleted contents of the VM from datastore datastore2 {{(pid=61907) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1779.751897] env[61907]: DEBUG nova.virt.vmwareapi.vmops [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Instance destroyed {{(pid=61907) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1779.752094] env[61907]: INFO nova.compute.manager [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1779.752353] env[61907]: DEBUG oslo.service.loopingcall [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61907) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1779.752558] env[61907]: DEBUG nova.compute.manager [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Deallocating network for instance {{(pid=61907) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2295}} [ 1779.752653] env[61907]: DEBUG nova.network.neutron [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] deallocate_for_instance() {{(pid=61907) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1780.204883] env[61907]: DEBUG nova.compute.manager [req-327e9e9b-ae1b-4544-a81c-9d2e017f8154 req-9f6e41e9-8c71-41d7-94b8-f85428eb4c73 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Received event network-vif-deleted-7528f86d-8b84-4a22-970f-f3db5232a72b {{(pid=61907) external_instance_event /opt/stack/nova/nova/compute/manager.py:11474}} [ 1780.205260] env[61907]: INFO nova.compute.manager [req-327e9e9b-ae1b-4544-a81c-9d2e017f8154 req-9f6e41e9-8c71-41d7-94b8-f85428eb4c73 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Neutron deleted interface 7528f86d-8b84-4a22-970f-f3db5232a72b; detaching it from the instance and deleting it from the info cache [ 1780.205260] env[61907]: DEBUG nova.network.neutron [req-327e9e9b-ae1b-4544-a81c-9d2e017f8154 req-9f6e41e9-8c71-41d7-94b8-f85428eb4c73 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1780.686269] env[61907]: DEBUG nova.network.neutron [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Updating instance_info_cache with network_info: [] {{(pid=61907) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1780.708125] env[61907]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf6df60a-a9ea-4b80-80bd-3008b6a9e527 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1780.718140] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e9b399-8e02-4162-9521-8469fea68262 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1780.738803] env[61907]: DEBUG nova.compute.manager [req-327e9e9b-ae1b-4544-a81c-9d2e017f8154 req-9f6e41e9-8c71-41d7-94b8-f85428eb4c73 service nova] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Detach interface failed, port_id=7528f86d-8b84-4a22-970f-f3db5232a72b, reason: Instance 5261031b-de16-432b-87f8-68c735995ce4 could not be found. {{(pid=61907) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11308}} [ 1781.189984] env[61907]: INFO nova.compute.manager [-] [instance: 5261031b-de16-432b-87f8-68c735995ce4] Took 1.44 seconds to deallocate network for instance. [ 1781.697751] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1781.698118] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1781.698270] env[61907]: DEBUG nova.objects.instance [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lazy-loading 'resources' on Instance uuid 5261031b-de16-432b-87f8-68c735995ce4 {{(pid=61907) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1782.232140] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c540d0-0478-4dc6-9995-60927f6d5c92 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1782.239614] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e3c292-ac7b-4ba4-b783-46ac19744348 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1782.269051] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc66b24a-df20-4ac6-a1b9-22cc06807e99 {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1782.275959] env[61907]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20b9a000-739d-441f-98bf-127600a67ccd {{(pid=61907) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1782.288447] env[61907]: DEBUG nova.compute.provider_tree [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed in ProviderTree for provider: 91bca385-a423-4ca4-9da0-aeb4615e22d3 {{(pid=61907) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1782.791060] env[61907]: DEBUG nova.scheduler.client.report [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Inventory has not changed for provider 91bca385-a423-4ca4-9da0-aeb4615e22d3 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 148, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61907) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:955}} [ 1783.297899] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1783.316741] env[61907]: INFO nova.scheduler.client.report [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Deleted allocations for instance 5261031b-de16-432b-87f8-68c735995ce4 [ 1783.824683] env[61907]: DEBUG oslo_concurrency.lockutils [None req-8f6bb261-9905-4cc4-8969-f1e6b2af2056 tempest-AttachVolumeNegativeTest-1137359015 tempest-AttachVolumeNegativeTest-1137359015-project-member] Lock "5261031b-de16-432b-87f8-68c735995ce4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.686s {{(pid=61907) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}